2024-12-05 15:11:49,982 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc 2024-12-05 15:11:49,994 main DEBUG Took 0.010315 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2024-12-05 15:11:49,994 main DEBUG PluginManager 'Core' found 129 plugins 2024-12-05 15:11:49,995 main DEBUG PluginManager 'Level' found 0 plugins 2024-12-05 15:11:49,995 main DEBUG PluginManager 'Lookup' found 16 plugins 2024-12-05 15:11:49,997 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,003 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2024-12-05 15:11:50,016 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,017 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,018 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,018 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,019 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,019 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,020 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,020 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,020 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,021 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,021 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,022 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,022 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,022 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,023 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,023 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,024 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,024 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,024 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,025 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,025 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,025 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,026 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,026 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2024-12-05 15:11:50,026 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,027 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2024-12-05 15:11:50,028 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2024-12-05 15:11:50,030 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2024-12-05 15:11:50,031 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2024-12-05 15:11:50,032 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2024-12-05 15:11:50,033 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2024-12-05 15:11:50,033 main DEBUG PluginManager 'Converter' found 47 plugins 2024-12-05 15:11:50,042 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2024-12-05 15:11:50,045 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2024-12-05 15:11:50,046 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2024-12-05 15:11:50,047 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2024-12-05 15:11:50,047 main DEBUG createAppenders(={Console}) 2024-12-05 15:11:50,048 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc initialized 2024-12-05 15:11:50,048 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc 2024-12-05 15:11:50,048 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@55de24cc OK. 2024-12-05 15:11:50,049 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2024-12-05 15:11:50,049 main DEBUG OutputStream closed 2024-12-05 15:11:50,049 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2024-12-05 15:11:50,050 main DEBUG Appender DefaultConsole-1 stopped with status true 2024-12-05 15:11:50,050 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@53ce1329 OK 2024-12-05 15:11:50,133 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2024-12-05 15:11:50,135 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2024-12-05 15:11:50,137 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2024-12-05 15:11:50,137 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2024-12-05 15:11:50,138 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2024-12-05 15:11:50,138 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2024-12-05 15:11:50,139 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2024-12-05 15:11:50,139 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2024-12-05 15:11:50,139 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2024-12-05 15:11:50,139 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2024-12-05 15:11:50,140 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2024-12-05 15:11:50,140 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2024-12-05 15:11:50,140 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2024-12-05 15:11:50,141 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2024-12-05 15:11:50,141 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2024-12-05 15:11:50,141 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2024-12-05 15:11:50,142 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2024-12-05 15:11:50,142 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2024-12-05 15:11:50,145 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-05 15:11:50,145 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-logging/target/hbase-logging-2.7.0-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@6dab9b6d) with optional ClassLoader: null 2024-12-05 15:11:50,145 main DEBUG Shutdown hook enabled. Registering a new one. 2024-12-05 15:11:50,146 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@6dab9b6d] started OK. 2024-12-05T15:11:50,456 DEBUG [main {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86 2024-12-05 15:11:50,460 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2024-12-05 15:11:50,461 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2024-12-05T15:11:50,474 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.regionserver.wal.TestLogRolling timeout: 13 mins 2024-12-05T15:11:50,509 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testSlowSyncLogRolling Thread=12, OpenFileDescriptor=286, MaxFileDescriptor=1048576, SystemLoadAverage=363, ProcessCount=11, AvailableMemoryMB=9665 2024-12-05T15:11:50,512 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T15:11:50,515 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef, deleteOnExit=true 2024-12-05T15:11:50,516 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-05T15:11:50,516 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/test.cache.data in system properties and HBase conf 2024-12-05T15:11:50,517 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T15:11:50,517 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir in system properties and HBase conf 2024-12-05T15:11:50,518 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T15:11:50,519 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T15:11:50,519 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-05T15:11:50,620 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2024-12-05T15:11:50,719 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T15:11:50,723 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:11:50,724 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:11:50,724 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T15:11:50,725 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:11:50,726 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T15:11:50,726 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T15:11:50,727 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:11:50,727 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:11:50,728 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T15:11:50,728 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/nfs.dump.dir in system properties and HBase conf 2024-12-05T15:11:50,729 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/java.io.tmpdir in system properties and HBase conf 2024-12-05T15:11:50,730 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:11:50,730 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T15:11:50,731 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T15:11:51,260 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:11:51,618 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2024-12-05T15:11:51,708 INFO [Time-limited test {}] log.Log(170): Logging initialized @2522ms to org.eclipse.jetty.util.log.Slf4jLog 2024-12-05T15:11:51,807 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:11:51,899 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:11:51,920 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:11:51,920 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:11:51,921 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:11:51,936 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:11:51,938 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@88aab13{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:11:51,939 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@74468826{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:11:52,165 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5682c4d1{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/java.io.tmpdir/jetty-localhost-39159-hadoop-hdfs-3_4_1-tests_jar-_-any-8465317095575108823/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:11:52,178 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1ff1a6c1{HTTP/1.1, (http/1.1)}{localhost:39159} 2024-12-05T15:11:52,179 INFO [Time-limited test {}] server.Server(415): Started @2994ms 2024-12-05T15:11:52,212 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:11:52,597 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:11:52,604 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:11:52,605 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:11:52,605 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:11:52,606 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:11:52,606 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3cf7922e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:11:52,607 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2276bd44{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:11:52,729 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5c4b1b4{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/java.io.tmpdir/jetty-localhost-36129-hadoop-hdfs-3_4_1-tests_jar-_-any-8547071436159583734/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:11:52,730 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@9fb4bbe{HTTP/1.1, (http/1.1)}{localhost:36129} 2024-12-05T15:11:52,730 INFO [Time-limited test {}] server.Server(415): Started @3545ms 2024-12-05T15:11:52,795 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:11:52,949 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:11:52,959 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:11:52,961 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:11:52,962 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:11:52,962 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:11:52,963 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@670df016{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:11:52,964 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4debea22{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:11:53,114 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2c482eac{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/java.io.tmpdir/jetty-localhost-36289-hadoop-hdfs-3_4_1-tests_jar-_-any-13106051344838851760/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:11:53,115 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@408e1d66{HTTP/1.1, (http/1.1)}{localhost:36289} 2024-12-05T15:11:53,115 INFO [Time-limited test {}] server.Server(415): Started @3931ms 2024-12-05T15:11:53,118 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:11:53,274 WARN [Thread-95 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data3/current/BP-1798444416-172.17.0.2-1733411511367/current, will proceed with Du for space computation calculation, 2024-12-05T15:11:53,274 WARN [Thread-97 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data1/current/BP-1798444416-172.17.0.2-1733411511367/current, will proceed with Du for space computation calculation, 2024-12-05T15:11:53,274 WARN [Thread-96 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data4/current/BP-1798444416-172.17.0.2-1733411511367/current, will proceed with Du for space computation calculation, 2024-12-05T15:11:53,275 WARN [Thread-98 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data2/current/BP-1798444416-172.17.0.2-1733411511367/current, will proceed with Du for space computation calculation, 2024-12-05T15:11:53,343 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:11:53,344 WARN [Thread-82 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:11:53,422 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7f90de3327972eb7 with lease ID 0x2089a971a0f6e7c0: Processing first storage report for DS-548361aa-2498-40fa-a485-f4a446ed184a from datanode DatanodeRegistration(127.0.0.1:40039, datanodeUuid=795230d8-3c86-4981-a198-c31c3ffefc5c, infoPort=40255, infoSecurePort=0, ipcPort=40315, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367) 2024-12-05T15:11:53,424 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7f90de3327972eb7 with lease ID 0x2089a971a0f6e7c0: from storage DS-548361aa-2498-40fa-a485-f4a446ed184a node DatanodeRegistration(127.0.0.1:40039, datanodeUuid=795230d8-3c86-4981-a198-c31c3ffefc5c, infoPort=40255, infoSecurePort=0, ipcPort=40315, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-05T15:11:53,424 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x186c2ab9db50e712 with lease ID 0x2089a971a0f6e7c1: Processing first storage report for DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51 from datanode DatanodeRegistration(127.0.0.1:46131, datanodeUuid=fb982a57-09ca-4495-ba58-d9a24c0e5bbe, infoPort=34571, infoSecurePort=0, ipcPort=36029, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367) 2024-12-05T15:11:53,424 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x186c2ab9db50e712 with lease ID 0x2089a971a0f6e7c1: from storage DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51 node DatanodeRegistration(127.0.0.1:46131, datanodeUuid=fb982a57-09ca-4495-ba58-d9a24c0e5bbe, infoPort=34571, infoSecurePort=0, ipcPort=36029, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:11:53,425 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7f90de3327972eb7 with lease ID 0x2089a971a0f6e7c0: Processing first storage report for DS-0650e0a5-994e-4d54-9d96-cf3b0f9f3193 from datanode DatanodeRegistration(127.0.0.1:40039, datanodeUuid=795230d8-3c86-4981-a198-c31c3ffefc5c, infoPort=40255, infoSecurePort=0, ipcPort=40315, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367) 2024-12-05T15:11:53,425 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7f90de3327972eb7 with lease ID 0x2089a971a0f6e7c0: from storage DS-0650e0a5-994e-4d54-9d96-cf3b0f9f3193 node DatanodeRegistration(127.0.0.1:40039, datanodeUuid=795230d8-3c86-4981-a198-c31c3ffefc5c, infoPort=40255, infoSecurePort=0, ipcPort=40315, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367), blocks: 0, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-05T15:11:53,425 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x186c2ab9db50e712 with lease ID 0x2089a971a0f6e7c1: Processing first storage report for DS-f8edd337-95be-4bc7-abef-b504d35d19ea from datanode DatanodeRegistration(127.0.0.1:46131, datanodeUuid=fb982a57-09ca-4495-ba58-d9a24c0e5bbe, infoPort=34571, infoSecurePort=0, ipcPort=36029, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367) 2024-12-05T15:11:53,425 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x186c2ab9db50e712 with lease ID 0x2089a971a0f6e7c1: from storage DS-f8edd337-95be-4bc7-abef-b504d35d19ea node DatanodeRegistration(127.0.0.1:46131, datanodeUuid=fb982a57-09ca-4495-ba58-d9a24c0e5bbe, infoPort=34571, infoSecurePort=0, ipcPort=36029, storageInfo=lv=-57;cid=testClusterID;nsid=1950772327;c=1733411511367), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:11:53,558 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86 2024-12-05T15:11:53,663 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/zookeeper_0, clientPort=52436, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T15:11:53,673 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=52436 2024-12-05T15:11:53,688 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:53,693 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:53,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:11:53,981 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:11:54,384 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b with version=8 2024-12-05T15:11:54,385 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1460): Setting hbase.fs.tmp.dir to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/hbase-staging 2024-12-05T15:11:54,509 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2024-12-05T15:11:54,779 INFO [Time-limited test {}] client.ConnectionUtils(129): master/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:11:54,799 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:11:54,799 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:11:54,799 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:11:54,799 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:11:54,800 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:11:54,934 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:11:54,994 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2024-12-05T15:11:55,004 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2024-12-05T15:11:55,008 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:11:55,036 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 29190 (auto-detected) 2024-12-05T15:11:55,037 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:02 (auto-detected) 2024-12-05T15:11:55,057 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:35771 2024-12-05T15:11:55,065 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:55,067 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:55,080 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:35771 connecting to ZooKeeper ensemble=127.0.0.1:52436 2024-12-05T15:11:55,113 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:357710x0, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:11:55,115 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:35771-0x10062bc5e100000 connected 2024-12-05T15:11:55,141 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:11:55,144 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:11:55,146 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:11:55,150 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35771 2024-12-05T15:11:55,151 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35771 2024-12-05T15:11:55,151 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35771 2024-12-05T15:11:55,152 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35771 2024-12-05T15:11:55,152 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35771 2024-12-05T15:11:55,159 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b, hbase.cluster.distributed=false 2024-12-05T15:11:55,230 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:11:55,230 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:11:55,230 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:11:55,230 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:11:55,231 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:11:55,231 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:11:55,233 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:11:55,235 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:11:55,236 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:37867 2024-12-05T15:11:55,238 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T15:11:55,244 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T15:11:55,246 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:55,253 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:55,259 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:37867 connecting to ZooKeeper ensemble=127.0.0.1:52436 2024-12-05T15:11:55,264 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:378670x0, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:11:55,264 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:378670x0, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:11:55,265 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37867-0x10062bc5e100001 connected 2024-12-05T15:11:55,267 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:11:55,268 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:11:55,273 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37867 2024-12-05T15:11:55,273 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37867 2024-12-05T15:11:55,276 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37867 2024-12-05T15:11:55,280 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37867 2024-12-05T15:11:55,280 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37867 2024-12-05T15:11:55,282 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:55,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:11:55,290 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:11:55,292 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:55,306 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;4ecb0f57b706:35771 2024-12-05T15:11:55,314 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:11:55,314 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:11:55,315 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,315 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,316 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:11:55,317 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:11:55,317 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/4ecb0f57b706,35771,1733411514503 from backup master directory 2024-12-05T15:11:55,321 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:55,321 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:11:55,321 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:11:55,322 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:11:55,322 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:55,325 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2024-12-05T15:11:55,326 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2024-12-05T15:11:55,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:11:55,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:11:55,404 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/hbase.id with ID: 97337b73-5af2-4742-b65e-ecfeae47f42a 2024-12-05T15:11:55,446 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:55,474 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,475 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:11:55,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:11:55,509 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:11:55,511 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T15:11:55,518 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:11:55,550 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:11:55,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:11:55,570 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store 2024-12-05T15:11:55,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:11:55,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:11:55,591 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2024-12-05T15:11:55,591 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:55,592 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:11:55,592 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:11:55,592 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:11:55,593 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:11:55,593 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:11:55,593 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:11:55,593 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:11:55,595 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/.initializing 2024-12-05T15:11:55,595 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/WALs/4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:55,612 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C35771%2C1733411514503, suffix=, logDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/WALs/4ecb0f57b706,35771,1733411514503, archiveDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/oldWALs, maxLogs=10 2024-12-05T15:11:55,620 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C35771%2C1733411514503.1733411515618 2024-12-05T15:11:55,621 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.CommonFSUtils$DfsBuilderUtility(752): Using builder API via reflection for DFS file creation replicate flag. 2024-12-05T15:11:55,621 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.CommonFSUtils$DfsBuilderUtility(762): Using builder API via reflection for DFS file creation noLocalWrite flag. 2024-12-05T15:11:55,646 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/WALs/4ecb0f57b706,35771,1733411514503/4ecb0f57b706%2C35771%2C1733411514503.1733411515618 2024-12-05T15:11:55,654 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:11:55,655 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:11:55,655 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:55,659 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,660 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,699 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,729 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T15:11:55,734 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:55,737 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:11:55,738 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,741 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T15:11:55,742 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:55,743 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:11:55,743 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,746 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T15:11:55,746 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:55,747 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:11:55,747 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,750 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T15:11:55,750 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:55,751 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:11:55,756 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,757 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,767 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T15:11:55,770 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:11:55,775 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:11:55,776 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=778799, jitterRate=-0.009705930948257446}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T15:11:55,780 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:11:55,782 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T15:11:55,813 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2b0af62f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:11:55,850 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-05T15:11:55,862 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T15:11:55,862 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T15:11:55,865 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T15:11:55,866 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 1 msec 2024-12-05T15:11:55,872 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 5 msec 2024-12-05T15:11:55,872 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T15:11:55,897 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T15:11:55,909 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T15:11:55,911 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-05T15:11:55,913 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T15:11:55,914 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T15:11:55,916 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-05T15:11:55,918 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T15:11:55,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T15:11:55,923 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-05T15:11:55,924 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T15:11:55,926 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T15:11:55,935 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T15:11:55,937 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T15:11:55,941 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:11:55,941 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:11:55,941 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,941 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,942 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=4ecb0f57b706,35771,1733411514503, sessionid=0x10062bc5e100000, setting cluster-up flag (Was=false) 2024-12-05T15:11:55,955 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,955 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,960 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T15:11:55,962 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:55,968 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,968 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:55,973 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T15:11:55,974 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:55,996 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;4ecb0f57b706:37867 2024-12-05T15:11:55,998 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1008): ClusterId : 97337b73-5af2-4742-b65e-ecfeae47f42a 2024-12-05T15:11:56,001 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T15:11:56,006 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T15:11:56,007 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T15:11:56,010 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T15:11:56,011 DEBUG [RS:0;4ecb0f57b706:37867 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@506a4e3f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:11:56,014 DEBUG [RS:0;4ecb0f57b706:37867 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@662bd158, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:11:56,018 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-05T15:11:56,018 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-05T15:11:56,018 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-05T15:11:56,021 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,35771,1733411514503 with isa=4ecb0f57b706/172.17.0.2:37867, startcode=1733411515229 2024-12-05T15:11:56,039 DEBUG [RS:0;4ecb0f57b706:37867 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:11:56,077 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-05T15:11:56,087 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-05T15:11:56,091 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T15:11:56,099 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 4ecb0f57b706,35771,1733411514503 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T15:11:56,104 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:11:56,104 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:11:56,105 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:11:56,105 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:11:56,105 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/4ecb0f57b706:0, corePoolSize=10, maxPoolSize=10 2024-12-05T15:11:56,105 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,106 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:11:56,106 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,111 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733411546110 2024-12-05T15:11:56,112 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46511, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:11:56,113 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T15:11:56,114 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T15:11:56,115 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:11:56,116 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-05T15:11:56,117 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T15:11:56,117 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T15:11:56,118 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T15:11:56,118 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T15:11:56,119 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,120 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T15:11:56,120 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:56,121 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:11:56,122 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T15:11:56,122 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T15:11:56,119 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35771 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3280) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-05T15:11:56,128 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T15:11:56,128 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T15:11:56,130 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411516130,5,FailOnTimeoutGroup] 2024-12-05T15:11:56,131 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411516130,5,FailOnTimeoutGroup] 2024-12-05T15:11:56,131 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,131 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T15:11:56,132 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,133 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:11:56,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:11:56,152 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-05T15:11:56,153 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b 2024-12-05T15:11:56,161 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(3097): Master is not running yet 2024-12-05T15:11:56,161 WARN [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2024-12-05T15:11:56,165 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:11:56,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:11:56,168 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:56,171 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:11:56,174 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:11:56,175 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:56,176 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:11:56,176 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:11:56,179 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:11:56,179 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:56,180 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:11:56,181 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:11:56,183 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:11:56,183 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:56,184 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:11:56,185 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740 2024-12-05T15:11:56,186 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740 2024-12-05T15:11:56,190 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:11:56,192 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:11:56,196 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:11:56,197 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=861625, jitterRate=0.09561312198638916}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:11:56,199 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:11:56,200 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:11:56,200 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:11:56,200 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:11:56,200 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:11:56,200 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:11:56,201 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:11:56,201 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:11:56,204 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:11:56,204 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-05T15:11:56,213 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T15:11:56,224 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T15:11:56,226 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T15:11:56,263 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,35771,1733411514503 with isa=4ecb0f57b706/172.17.0.2:37867, startcode=1733411515229 2024-12-05T15:11:56,265 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35771 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,267 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=35771 {}] master.ServerManager(486): Registering regionserver=4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,275 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b 2024-12-05T15:11:56,276 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:35505 2024-12-05T15:11:56,276 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-05T15:11:56,281 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:11:56,282 DEBUG [RS:0;4ecb0f57b706:37867 {}] zookeeper.ZKUtil(111): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,282 WARN [RS:0;4ecb0f57b706:37867 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:11:56,282 INFO [RS:0;4ecb0f57b706:37867 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:11:56,282 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,285 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [4ecb0f57b706,37867,1733411515229] 2024-12-05T15:11:56,299 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-05T15:11:56,313 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T15:11:56,324 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T15:11:56,327 INFO [RS:0;4ecb0f57b706:37867 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:11:56,327 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,328 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-05T15:11:56,335 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,335 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,336 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,336 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,336 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,336 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,336 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:11:56,336 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,337 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,337 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,337 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,337 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:11:56,337 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:11:56,337 DEBUG [RS:0;4ecb0f57b706:37867 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:11:56,339 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,339 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,340 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,340 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,340 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,37867,1733411515229-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:11:56,363 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T15:11:56,366 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,37867,1733411515229-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:56,377 WARN [4ecb0f57b706:35771 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-05T15:11:56,391 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.Replication(204): 4ecb0f57b706,37867,1733411515229 started 2024-12-05T15:11:56,391 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1767): Serving as 4ecb0f57b706,37867,1733411515229, RpcServer on 4ecb0f57b706/172.17.0.2:37867, sessionid=0x10062bc5e100001 2024-12-05T15:11:56,392 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T15:11:56,392 DEBUG [RS:0;4ecb0f57b706:37867 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,392 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,37867,1733411515229' 2024-12-05T15:11:56,392 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T15:11:56,393 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T15:11:56,393 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T15:11:56,394 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T15:11:56,394 DEBUG [RS:0;4ecb0f57b706:37867 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,394 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,37867,1733411515229' 2024-12-05T15:11:56,394 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T15:11:56,394 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T15:11:56,395 DEBUG [RS:0;4ecb0f57b706:37867 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T15:11:56,395 INFO [RS:0;4ecb0f57b706:37867 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T15:11:56,395 INFO [RS:0;4ecb0f57b706:37867 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T15:11:56,503 INFO [RS:0;4ecb0f57b706:37867 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C37867%2C1733411515229, suffix=, logDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229, archiveDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs, maxLogs=32 2024-12-05T15:11:56,506 INFO [RS:0;4ecb0f57b706:37867 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411516506 2024-12-05T15:11:56,515 INFO [RS:0;4ecb0f57b706:37867 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411516506 2024-12-05T15:11:56,515 DEBUG [RS:0;4ecb0f57b706:37867 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:11:56,629 DEBUG [4ecb0f57b706:35771 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-05T15:11:56,633 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,638 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,37867,1733411515229, state=OPENING 2024-12-05T15:11:56,643 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T15:11:56,645 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:56,645 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:56,646 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:11:56,646 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:11:56,648 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=4ecb0f57b706,37867,1733411515229}] 2024-12-05T15:11:56,823 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,824 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T15:11:56,828 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59086, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T15:11:56,839 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-05T15:11:56,840 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:11:56,844 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C37867%2C1733411515229.meta, suffix=.meta, logDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229, archiveDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs, maxLogs=32 2024-12-05T15:11:56,846 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.meta.1733411516845.meta 2024-12-05T15:11:56,854 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.meta.1733411516845.meta 2024-12-05T15:11:56,855 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:11:56,855 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:11:56,856 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T15:11:56,915 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T15:11:56,920 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T15:11:56,924 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T15:11:56,925 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:56,925 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-05T15:11:56,925 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-05T15:11:56,928 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:11:56,930 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:11:56,930 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:56,931 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:11:56,931 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:11:56,932 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:11:56,932 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:56,933 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:11:56,933 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:11:56,934 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:11:56,935 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:56,935 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:11:56,937 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740 2024-12-05T15:11:56,939 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740 2024-12-05T15:11:56,941 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:11:56,944 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:11:56,946 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=830519, jitterRate=0.05606013536453247}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:11:56,947 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:11:56,954 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733411516817 2024-12-05T15:11:56,964 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T15:11:56,965 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-05T15:11:56,966 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:56,967 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,37867,1733411515229, state=OPEN 2024-12-05T15:11:56,973 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:11:56,973 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:11:56,973 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:11:56,973 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:11:56,977 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T15:11:56,977 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=4ecb0f57b706,37867,1733411515229 in 325 msec 2024-12-05T15:11:56,983 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T15:11:56,983 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 767 msec 2024-12-05T15:11:56,988 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 974 msec 2024-12-05T15:11:56,989 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733411516988, completionTime=-1 2024-12-05T15:11:56,989 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-05T15:11:56,989 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-05T15:11:57,027 DEBUG [hconnection-0x473760c6-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:11:57,030 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59090, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:11:57,040 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-05T15:11:57,040 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733411577040 2024-12-05T15:11:57,040 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733411637040 2024-12-05T15:11:57,040 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 51 msec 2024-12-05T15:11:57,069 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35771,1733411514503-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:57,069 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35771,1733411514503-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:57,069 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35771,1733411514503-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:57,071 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-4ecb0f57b706:35771, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:57,072 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:57,077 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-05T15:11:57,081 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-05T15:11:57,083 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:11:57,090 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-05T15:11:57,093 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:11:57,095 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:57,097 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:11:57,109 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:11:57,109 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:11:57,112 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 2ffa63508ac81c218f03e91221ac2e05, NAME => 'hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b 2024-12-05T15:11:57,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:11:57,123 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:11:57,124 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:57,124 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 2ffa63508ac81c218f03e91221ac2e05, disabling compactions & flushes 2024-12-05T15:11:57,124 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:11:57,124 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:11:57,124 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. after waiting 0 ms 2024-12-05T15:11:57,124 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:11:57,124 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:11:57,124 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 2ffa63508ac81c218f03e91221ac2e05: 2024-12-05T15:11:57,126 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:11:57,133 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733411517128"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411517128"}]},"ts":"1733411517128"} 2024-12-05T15:11:57,159 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:11:57,161 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:11:57,164 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411517161"}]},"ts":"1733411517161"} 2024-12-05T15:11:57,170 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-05T15:11:57,177 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=2ffa63508ac81c218f03e91221ac2e05, ASSIGN}] 2024-12-05T15:11:57,180 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=2ffa63508ac81c218f03e91221ac2e05, ASSIGN 2024-12-05T15:11:57,182 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=2ffa63508ac81c218f03e91221ac2e05, ASSIGN; state=OFFLINE, location=4ecb0f57b706,37867,1733411515229; forceNewPlan=false, retain=false 2024-12-05T15:11:57,333 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=2ffa63508ac81c218f03e91221ac2e05, regionState=OPENING, regionLocation=4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:57,337 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 2ffa63508ac81c218f03e91221ac2e05, server=4ecb0f57b706,37867,1733411515229}] 2024-12-05T15:11:57,492 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:57,499 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:11:57,499 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 2ffa63508ac81c218f03e91221ac2e05, NAME => 'hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:11:57,500 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:11:57,500 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:57,500 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:11:57,500 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:11:57,503 INFO [StoreOpener-2ffa63508ac81c218f03e91221ac2e05-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:11:57,506 INFO [StoreOpener-2ffa63508ac81c218f03e91221ac2e05-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 2ffa63508ac81c218f03e91221ac2e05 columnFamilyName info 2024-12-05T15:11:57,506 DEBUG [StoreOpener-2ffa63508ac81c218f03e91221ac2e05-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:57,507 INFO [StoreOpener-2ffa63508ac81c218f03e91221ac2e05-1 {}] regionserver.HStore(327): Store=2ffa63508ac81c218f03e91221ac2e05/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:11:57,509 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:11:57,509 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:11:57,513 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:11:57,517 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:11:57,518 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 2ffa63508ac81c218f03e91221ac2e05; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=842268, jitterRate=0.07099993526935577}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:11:57,519 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 2ffa63508ac81c218f03e91221ac2e05: 2024-12-05T15:11:57,521 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05., pid=6, masterSystemTime=1733411517492 2024-12-05T15:11:57,525 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:11:57,525 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:11:57,526 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=2ffa63508ac81c218f03e91221ac2e05, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:57,534 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T15:11:57,534 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 2ffa63508ac81c218f03e91221ac2e05, server=4ecb0f57b706,37867,1733411515229 in 193 msec 2024-12-05T15:11:57,538 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T15:11:57,538 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=2ffa63508ac81c218f03e91221ac2e05, ASSIGN in 357 msec 2024-12-05T15:11:57,539 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:11:57,540 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411517539"}]},"ts":"1733411517539"} 2024-12-05T15:11:57,543 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-05T15:11:57,546 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:11:57,549 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 463 msec 2024-12-05T15:11:57,594 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-05T15:11:57,596 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:11:57,596 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:57,596 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:11:57,630 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-05T15:11:57,647 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:11:57,652 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 27 msec 2024-12-05T15:11:57,665 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-05T15:11:57,678 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:11:57,683 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 16 msec 2024-12-05T15:11:57,692 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-05T15:11:57,695 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-05T15:11:57,696 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 2.373sec 2024-12-05T15:11:57,697 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T15:11:57,698 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T15:11:57,699 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T15:11:57,700 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T15:11:57,700 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T15:11:57,701 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35771,1733411514503-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:11:57,701 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35771,1733411514503-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T15:11:57,708 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-05T15:11:57,709 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T15:11:57,709 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35771,1733411514503-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:11:57,799 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x778ff29f to 127.0.0.1:52436 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@50389d05 2024-12-05T15:11:57,800 WARN [Time-limited test {}] client.ZKConnectionRegistry(90): ZKConnectionRegistry is deprecated. See https://hbase.apache.org/book.html#client.rpcconnectionregistry 2024-12-05T15:11:57,807 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@588da4fe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:11:57,810 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2024-12-05T15:11:57,810 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2024-12-05T15:11:57,820 DEBUG [hconnection-0x7dc6e43c-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:11:57,853 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:59106, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:11:57,862 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=4ecb0f57b706,35771,1733411514503 2024-12-05T15:11:57,863 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:11:57,871 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-05T15:11:57,878 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-05T15:11:57,881 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50346, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-05T15:11:57,887 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35771 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-05T15:11:57,888 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35771 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-05T15:11:57,891 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35771 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testSlowSyncLogRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:11:57,894 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35771 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling 2024-12-05T15:11:57,896 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:11:57,896 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35771 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testSlowSyncLogRolling" procId is: 9 2024-12-05T15:11:57,896 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:57,898 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:11:57,904 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35771 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:11:57,909 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741837_1013 (size=389) 2024-12-05T15:11:57,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741837_1013 (size=389) 2024-12-05T15:11:57,913 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => e2d67f93c8081f8f26afcb8999bd38fe, NAME => 'TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testSlowSyncLogRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b 2024-12-05T15:11:57,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741838_1014 (size=72) 2024-12-05T15:11:57,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741838_1014 (size=72) 2024-12-05T15:11:57,923 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:57,924 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1681): Closing e2d67f93c8081f8f26afcb8999bd38fe, disabling compactions & flushes 2024-12-05T15:11:57,924 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:11:57,924 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:11:57,924 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. after waiting 0 ms 2024-12-05T15:11:57,924 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:11:57,924 INFO [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:11:57,924 DEBUG [RegionOpenAndInit-TestLogRolling-testSlowSyncLogRolling-pool-0 {}] regionserver.HRegion(1635): Region close journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:11:57,926 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:11:57,926 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.","families":{"info":[{"qualifier":"regioninfo","vlen":71,"tag":[],"timestamp":"1733411517926"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411517926"}]},"ts":"1733411517926"} 2024-12-05T15:11:57,930 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:11:57,931 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:11:57,932 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testSlowSyncLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411517931"}]},"ts":"1733411517931"} 2024-12-05T15:11:57,935 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testSlowSyncLogRolling, state=ENABLING in hbase:meta 2024-12-05T15:11:57,940 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=e2d67f93c8081f8f26afcb8999bd38fe, ASSIGN}] 2024-12-05T15:11:57,942 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=e2d67f93c8081f8f26afcb8999bd38fe, ASSIGN 2024-12-05T15:11:57,943 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=e2d67f93c8081f8f26afcb8999bd38fe, ASSIGN; state=OFFLINE, location=4ecb0f57b706,37867,1733411515229; forceNewPlan=false, retain=false 2024-12-05T15:11:58,094 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=e2d67f93c8081f8f26afcb8999bd38fe, regionState=OPENING, regionLocation=4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:58,097 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure e2d67f93c8081f8f26afcb8999bd38fe, server=4ecb0f57b706,37867,1733411515229}] 2024-12-05T15:11:58,251 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:58,259 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:11:58,259 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => e2d67f93c8081f8f26afcb8999bd38fe, NAME => 'TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:11:58,259 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testSlowSyncLogRolling e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:11:58,259 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:11:58,259 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:11:58,260 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:11:58,262 INFO [StoreOpener-e2d67f93c8081f8f26afcb8999bd38fe-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:11:58,265 INFO [StoreOpener-e2d67f93c8081f8f26afcb8999bd38fe-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region e2d67f93c8081f8f26afcb8999bd38fe columnFamilyName info 2024-12-05T15:11:58,265 DEBUG [StoreOpener-e2d67f93c8081f8f26afcb8999bd38fe-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:11:58,266 INFO [StoreOpener-e2d67f93c8081f8f26afcb8999bd38fe-1 {}] regionserver.HStore(327): Store=e2d67f93c8081f8f26afcb8999bd38fe/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:11:58,267 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:11:58,268 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:11:58,271 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:11:58,275 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:11:58,276 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened e2d67f93c8081f8f26afcb8999bd38fe; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=701263, jitterRate=-0.10829856991767883}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:11:58,277 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:11:58,279 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe., pid=11, masterSystemTime=1733411518251 2024-12-05T15:11:58,282 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:11:58,282 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:11:58,283 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=e2d67f93c8081f8f26afcb8999bd38fe, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,37867,1733411515229 2024-12-05T15:11:58,290 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-05T15:11:58,290 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure e2d67f93c8081f8f26afcb8999bd38fe, server=4ecb0f57b706,37867,1733411515229 in 189 msec 2024-12-05T15:11:58,293 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-05T15:11:58,293 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testSlowSyncLogRolling, region=e2d67f93c8081f8f26afcb8999bd38fe, ASSIGN in 350 msec 2024-12-05T15:11:58,295 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:11:58,295 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testSlowSyncLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411518295"}]},"ts":"1733411518295"} 2024-12-05T15:11:58,298 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testSlowSyncLogRolling, state=ENABLED in hbase:meta 2024-12-05T15:11:58,301 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:11:58,304 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testSlowSyncLogRolling in 409 msec 2024-12-05T15:12:02,429 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-hbase.properties,hadoop-metrics2.properties 2024-12-05T15:12:02,473 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-05T15:12:02,474 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-05T15:12:02,475 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testSlowSyncLogRolling' 2024-12-05T15:12:04,991 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2024-12-05T15:12:04,991 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2024-12-05T15:12:04,993 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling 2024-12-05T15:12:04,993 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling Metrics about Tables on a single HBase RegionServer 2024-12-05T15:12:04,994 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-05T15:12:04,994 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2024-12-05T15:12:04,995 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T15:12:04,995 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2024-12-05T15:12:04,996 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-05T15:12:04,996 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-05T15:12:07,908 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=35771 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:12:07,909 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testSlowSyncLogRolling, procId: 9 completed 2024-12-05T15:12:07,913 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testSlowSyncLogRolling 2024-12-05T15:12:07,914 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:12:07,915 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411527914 2024-12-05T15:12:07,925 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411516506 with entries=4, filesize=947 B; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411527914 2024-12-05T15:12:07,926 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:12:07,926 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411516506 is not closed yet, will try archiving it next time 2024-12-05T15:12:07,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741833_1009 (size=955) 2024-12-05T15:12:07,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741833_1009 (size=955) 2024-12-05T15:12:19,956 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37867 {}] regionserver.HRegion(8581): Flush requested on e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:12:19,957 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing e2d67f93c8081f8f26afcb8999bd38fe 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:12:20,041 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/7a0ab1f09fd9498e999cff4a1ca04f23 is 1080, key is row0001/info:/1733411527932/Put/seqid=0 2024-12-05T15:12:20,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741840_1016 (size=12509) 2024-12-05T15:12:20,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741840_1016 (size=12509) 2024-12-05T15:12:20,056 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/7a0ab1f09fd9498e999cff4a1ca04f23 2024-12-05T15:12:20,117 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/7a0ab1f09fd9498e999cff4a1ca04f23 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23 2024-12-05T15:12:20,128 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23, entries=7, sequenceid=11, filesize=12.2 K 2024-12-05T15:12:20,132 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for e2d67f93c8081f8f26afcb8999bd38fe in 175ms, sequenceid=11, compaction requested=false 2024-12-05T15:12:20,132 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:12:23,563 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:12:27,968 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411547968 2024-12-05T15:12:28,178 INFO [Time-limited test {}] wal.AbstractFSWAL(1183): Slow sync cost: 206 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:12:28,180 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411527914 with entries=12, filesize=12.10 KB; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411547968 2024-12-05T15:12:28,180 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34571:34571),(127.0.0.1/127.0.0.1:40255:40255)] 2024-12-05T15:12:28,180 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411527914 is not closed yet, will try archiving it next time 2024-12-05T15:12:28,182 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741839_1015 (size=12399) 2024-12-05T15:12:28,183 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741839_1015 (size=12399) 2024-12-05T15:12:28,383 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:30,015 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:12:30,017 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:43852, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:12:30,586 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:32,790 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:34,994 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:34,994 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37867 {}] regionserver.HRegion(8581): Flush requested on e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:12:34,994 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing e2d67f93c8081f8f26afcb8999bd38fe 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:12:35,195 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:35,202 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/9fb9c3efa77f47768579b2ccfc9b804c is 1080, key is row0008/info:/1733411541958/Put/seqid=0 2024-12-05T15:12:35,214 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741842_1018 (size=12509) 2024-12-05T15:12:35,215 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741842_1018 (size=12509) 2024-12-05T15:12:35,215 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=21 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/9fb9c3efa77f47768579b2ccfc9b804c 2024-12-05T15:12:35,225 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/9fb9c3efa77f47768579b2ccfc9b804c as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/9fb9c3efa77f47768579b2ccfc9b804c 2024-12-05T15:12:35,236 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/9fb9c3efa77f47768579b2ccfc9b804c, entries=7, sequenceid=21, filesize=12.2 K 2024-12-05T15:12:35,437 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:35,438 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for e2d67f93c8081f8f26afcb8999bd38fe in 443ms, sequenceid=21, compaction requested=false 2024-12-05T15:12:35,438 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:12:35,438 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=24.4 K, sizeToCheck=16.0 K 2024-12-05T15:12:35,438 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:12:35,439 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23 because midkey is the same as first or last row 2024-12-05T15:12:37,197 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:37,711 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-05T15:12:37,711 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-05T15:12:39,400 WARN [sync.1 {}] wal.AbstractFSWAL(1346): Requesting log roll because we exceeded slow sync threshold; count=7, threshold=5, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:39,402 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C37867%2C1733411515229:(num 1733411547968) roll requested 2024-12-05T15:12:39,402 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 202 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:39,402 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411559402 2024-12-05T15:12:39,610 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(1183): Slow sync cost: 206 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:39,811 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:12:39,812 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411547968 with entries=8, filesize=7.55 KB; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411559402 2024-12-05T15:12:39,812 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:12:39,812 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411547968 is not closed yet, will try archiving it next time 2024-12-05T15:12:39,813 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411527914 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs/4ecb0f57b706%2C37867%2C1733411515229.1733411527914 2024-12-05T15:12:39,815 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741841_1017 (size=7739) 2024-12-05T15:12:39,815 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741841_1017 (size=7739) 2024-12-05T15:12:41,604 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:12:43,260 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region e2d67f93c8081f8f26afcb8999bd38fe, had cached 0 bytes from a total of 25018 2024-12-05T15:12:43,807 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 200 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:12:46,010 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:12:48,214 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 201 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:12:50,216 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [10,000] milli-secs(wait.for.ratio=[1]) 2024-12-05T15:12:50,216 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411570216 2024-12-05T15:12:53,563 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:12:55,228 INFO [Time-limited test {}] wal.AbstractFSWAL(1183): Slow sync cost: 5008 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:12:55,228 WARN [Time-limited test {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5008 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:12:55,228 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C37867%2C1733411515229:(num 1733411570216) roll requested 2024-12-05T15:12:57,072 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 2ffa63508ac81c218f03e91221ac2e05 changed from -1.0 to 0.0, refreshing cache 2024-12-05T15:13:00,229 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:00,229 WARN [sync.2 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:00,229 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411559402 with entries=4, filesize=4.63 KB; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411570216 2024-12-05T15:13:00,230 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:13:00,230 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411559402 is not closed yet, will try archiving it next time 2024-12-05T15:13:00,230 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411580230 2024-12-05T15:13:00,232 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741843_1019 (size=4753) 2024-12-05T15:13:00,233 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741843_1019 (size=4753) 2024-12-05T15:13:05,233 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5001 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:05,233 WARN [sync.3 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5001 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:05,233 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37867 {}] regionserver.HRegion(8581): Flush requested on e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:13:05,233 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing e2d67f93c8081f8f26afcb8999bd38fe 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:13:05,239 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(1183): Slow sync cost: 5005 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:05,239 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5005 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:07,234 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [10,000] milli-secs(wait.for.ratio=[1]) 2024-12-05T15:13:10,235 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:10,235 WARN [sync.4 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:10,239 INFO [sync.0 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:10,239 WARN [sync.0 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK], DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK]] 2024-12-05T15:13:10,240 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/5484fedef4824de38ed42183fa1dcc97 is 1080, key is row0015/info:/1733411556995/Put/seqid=0 2024-12-05T15:13:10,240 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411570216 with entries=2, filesize=1.52 KB; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411580230 2024-12-05T15:13:10,240 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34571:34571),(127.0.0.1/127.0.0.1:40255:40255)] 2024-12-05T15:13:10,241 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411570216 is not closed yet, will try archiving it next time 2024-12-05T15:13:10,241 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C37867%2C1733411515229:(num 1733411580230) roll requested 2024-12-05T15:13:10,241 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411590241 2024-12-05T15:13:10,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741844_1020 (size=1569) 2024-12-05T15:13:10,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741844_1020 (size=1569) 2024-12-05T15:13:10,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741846_1022 (size=12509) 2024-12-05T15:13:10,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741846_1022 (size=12509) 2024-12-05T15:13:10,254 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=31 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/5484fedef4824de38ed42183fa1dcc97 2024-12-05T15:13:10,264 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/5484fedef4824de38ed42183fa1dcc97 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/5484fedef4824de38ed42183fa1dcc97 2024-12-05T15:13:10,272 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/5484fedef4824de38ed42183fa1dcc97, entries=7, sequenceid=31, filesize=12.2 K 2024-12-05T15:13:15,249 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(1183): Slow sync cost: 5005 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:15,249 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5005 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:15,273 INFO [sync.1 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:15,273 WARN [sync.1 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:15,273 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for e2d67f93c8081f8f26afcb8999bd38fe in 10040ms, sequenceid=31, compaction requested=true 2024-12-05T15:13:15,273 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:13:15,274 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=36.6 K, sizeToCheck=16.0 K 2024-12-05T15:13:15,274 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:13:15,274 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23 because midkey is the same as first or last row 2024-12-05T15:13:15,275 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store e2d67f93c8081f8f26afcb8999bd38fe:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:13:15,275 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:13:15,276 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:13:15,278 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 37527 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:13:15,280 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.HStore(1540): e2d67f93c8081f8f26afcb8999bd38fe/info is initiating minor compaction (all files) 2024-12-05T15:13:15,280 INFO [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of e2d67f93c8081f8f26afcb8999bd38fe/info in TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:13:15,280 INFO [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23, hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/9fb9c3efa77f47768579b2ccfc9b804c, hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/5484fedef4824de38ed42183fa1dcc97] into tmpdir=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp, totalSize=36.6 K 2024-12-05T15:13:15,281 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] compactions.Compactor(224): Compacting 7a0ab1f09fd9498e999cff4a1ca04f23, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=11, earliestPutTs=1733411527932 2024-12-05T15:13:15,282 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] compactions.Compactor(224): Compacting 9fb9c3efa77f47768579b2ccfc9b804c, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=21, earliestPutTs=1733411541958 2024-12-05T15:13:15,282 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] compactions.Compactor(224): Compacting 5484fedef4824de38ed42183fa1dcc97, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=31, earliestPutTs=1733411556995 2024-12-05T15:13:15,311 INFO [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): e2d67f93c8081f8f26afcb8999bd38fe#info#compaction#3 average throughput is 10.77 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:13:15,312 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/de41700247d5458385bcd6a89f565fb5 is 1080, key is row0001/info:/1733411527932/Put/seqid=0 2024-12-05T15:13:15,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741848_1024 (size=27710) 2024-12-05T15:13:15,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741848_1024 (size=27710) 2024-12-05T15:13:15,328 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/de41700247d5458385bcd6a89f565fb5 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/de41700247d5458385bcd6a89f565fb5 2024-12-05T15:13:20,250 INFO [sync.2 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:20,250 WARN [sync.2 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:20,250 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411580230 with entries=1, filesize=430 B; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411590241 2024-12-05T15:13:20,250 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34571:34571),(127.0.0.1/127.0.0.1:40255:40255)] 2024-12-05T15:13:20,250 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411580230 is not closed yet, will try archiving it next time 2024-12-05T15:13:20,251 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411547968 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs/4ecb0f57b706%2C37867%2C1733411515229.1733411547968 2024-12-05T15:13:20,251 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C37867%2C1733411515229:(num 1733411600251) roll requested 2024-12-05T15:13:20,251 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411600251 2024-12-05T15:13:20,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741845_1021 (size=438) 2024-12-05T15:13:20,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741845_1021 (size=438) 2024-12-05T15:13:20,254 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411559402 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs/4ecb0f57b706%2C37867%2C1733411515229.1733411559402 2024-12-05T15:13:20,255 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411570216 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs/4ecb0f57b706%2C37867%2C1733411515229.1733411570216 2024-12-05T15:13:20,257 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411580230 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs/4ecb0f57b706%2C37867%2C1733411515229.1733411580230 2024-12-05T15:13:23,563 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:13:25,251 INFO [sync.3 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:25,252 WARN [sync.3 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:25,253 INFO [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in e2d67f93c8081f8f26afcb8999bd38fe/info of e2d67f93c8081f8f26afcb8999bd38fe into de41700247d5458385bcd6a89f565fb5(size=27.1 K), total size for store is 27.1 K. This selection was in queue for 0sec, and took 9sec to execute. 2024-12-05T15:13:25,253 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:13:25,254 INFO [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe., storeName=e2d67f93c8081f8f26afcb8999bd38fe/info, priority=13, startTime=1733411595275; duration=9sec 2024-12-05T15:13:25,254 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=27.1 K, sizeToCheck=16.0 K 2024-12-05T15:13:25,254 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:13:25,254 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/de41700247d5458385bcd6a89f565fb5 because midkey is the same as first or last row 2024-12-05T15:13:25,254 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:13:25,254 DEBUG [RS:0;4ecb0f57b706:37867-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: e2d67f93c8081f8f26afcb8999bd38fe:info 2024-12-05T15:13:25,260 INFO [sync.4 {}] wal.AbstractFSWAL(1183): Slow sync cost: 5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:25,260 WARN [sync.4 {}] wal.AbstractFSWAL(1189): Requesting log roll because we exceeded slow sync threshold; time=5000 ms, threshold=5000 ms, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:46131,DS-3e9adeb7-dda3-4821-93f1-f82a26ff3a51,DISK], DatanodeInfoWithStorage[127.0.0.1:40039,DS-548361aa-2498-40fa-a485-f4a446ed184a,DISK]] 2024-12-05T15:13:25,261 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411590241 with entries=1, filesize=531 B; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411600251 2024-12-05T15:13:25,261 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:13:25,261 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411590241 is not closed yet, will try archiving it next time 2024-12-05T15:13:25,262 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411605261 2024-12-05T15:13:25,264 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741847_1023 (size=539) 2024-12-05T15:13:25,269 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411590241 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs/4ecb0f57b706%2C37867%2C1733411515229.1733411590241 2024-12-05T15:13:25,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741847_1023 (size=539) 2024-12-05T15:13:25,296 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411600251 with entries=1, filesize=1.22 KB; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411605261 2024-12-05T15:13:25,296 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:13:25,296 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411600251 is not closed yet, will try archiving it next time 2024-12-05T15:13:25,296 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C37867%2C1733411515229:(num 1733411605261) roll requested 2024-12-05T15:13:25,297 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C37867%2C1733411515229.1733411605296 2024-12-05T15:13:25,300 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741849_1025 (size=1258) 2024-12-05T15:13:25,300 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741849_1025 (size=1258) 2024-12-05T15:13:25,324 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411605261 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411605296 2024-12-05T15:13:25,324 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40255:40255),(127.0.0.1/127.0.0.1:34571:34571)] 2024-12-05T15:13:25,324 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411605261 is not closed yet, will try archiving it next time 2024-12-05T15:13:25,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741850_1026 (size=93) 2024-12-05T15:13:25,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741850_1026 (size=93) 2024-12-05T15:13:25,734 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229/4ecb0f57b706%2C37867%2C1733411515229.1733411605261 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs/4ecb0f57b706%2C37867%2C1733411515229.1733411605261 2024-12-05T15:13:28,260 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region e2d67f93c8081f8f26afcb8999bd38fe, had cached 0 bytes from a total of 27710 2024-12-05T15:13:37,284 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37867 {}] regionserver.HRegion(8581): Flush requested on e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:13:37,284 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing e2d67f93c8081f8f26afcb8999bd38fe 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:13:37,293 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/a1b3ecf97b9144f2af4a0df404574078 is 1080, key is row0022/info:/1733411605263/Put/seqid=0 2024-12-05T15:13:37,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741852_1028 (size=12509) 2024-12-05T15:13:37,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741852_1028 (size=12509) 2024-12-05T15:13:37,722 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=42 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/a1b3ecf97b9144f2af4a0df404574078 2024-12-05T15:13:37,731 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/a1b3ecf97b9144f2af4a0df404574078 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/a1b3ecf97b9144f2af4a0df404574078 2024-12-05T15:13:37,740 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/a1b3ecf97b9144f2af4a0df404574078, entries=7, sequenceid=42, filesize=12.2 K 2024-12-05T15:13:37,741 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for e2d67f93c8081f8f26afcb8999bd38fe in 457ms, sequenceid=42, compaction requested=false 2024-12-05T15:13:37,741 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:13:37,742 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=39.3 K, sizeToCheck=16.0 K 2024-12-05T15:13:37,742 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:13:37,742 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/de41700247d5458385bcd6a89f565fb5 because midkey is the same as first or last row 2024-12-05T15:13:45,294 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-05T15:13:45,295 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-05T15:13:45,295 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x778ff29f to 127.0.0.1:52436 2024-12-05T15:13:45,295 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:13:45,296 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T15:13:45,296 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=312489965, stopped=false 2024-12-05T15:13:45,296 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=4ecb0f57b706,35771,1733411514503 2024-12-05T15:13:45,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:13:45,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:13:45,298 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-05T15:13:45,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:45,298 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:45,298 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:13:45,298 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,37867,1733411515229' ***** 2024-12-05T15:13:45,299 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-05T15:13:45,299 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:13:45,299 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:13:45,299 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T15:13:45,299 INFO [RS:0;4ecb0f57b706:37867 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T15:13:45,299 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-05T15:13:45,299 INFO [RS:0;4ecb0f57b706:37867 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T15:13:45,300 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(3579): Received CLOSE for e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:13:45,300 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(3579): Received CLOSE for 2ffa63508ac81c218f03e91221ac2e05 2024-12-05T15:13:45,300 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,37867,1733411515229 2024-12-05T15:13:45,300 DEBUG [RS:0;4ecb0f57b706:37867 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:13:45,300 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T15:13:45,300 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T15:13:45,300 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T15:13:45,301 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing e2d67f93c8081f8f26afcb8999bd38fe, disabling compactions & flushes 2024-12-05T15:13:45,301 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-05T15:13:45,301 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:13:45,301 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1603): Online Regions={e2d67f93c8081f8f26afcb8999bd38fe=TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe., 1588230740=hbase:meta,,1.1588230740, 2ffa63508ac81c218f03e91221ac2e05=hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05.} 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. after waiting 0 ms 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:13:45,301 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:13:45,301 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing e2d67f93c8081f8f26afcb8999bd38fe 1/1 column families, dataSize=3.15 KB heapSize=3.63 KB 2024-12-05T15:13:45,301 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:13:45,301 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.81 KB heapSize=5.32 KB 2024-12-05T15:13:45,301 DEBUG [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 2ffa63508ac81c218f03e91221ac2e05, e2d67f93c8081f8f26afcb8999bd38fe 2024-12-05T15:13:45,307 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/f8472c8f07634d0a95e9015d4c4ef00e is 1080, key is row0029/info:/1733411619286/Put/seqid=0 2024-12-05T15:13:45,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741853_1029 (size=8193) 2024-12-05T15:13:45,316 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741853_1029 (size=8193) 2024-12-05T15:13:45,316 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=3.15 KB at sequenceid=48 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/f8472c8f07634d0a95e9015d4c4ef00e 2024-12-05T15:13:45,325 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/.tmp/info/f8472c8f07634d0a95e9015d4c4ef00e as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/f8472c8f07634d0a95e9015d4c4ef00e 2024-12-05T15:13:45,326 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/.tmp/info/ac73eb4532ad40ea8f4c4c5f29e8da79 is 195, key is TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe./info:regioninfo/1733411518283/Put/seqid=0 2024-12-05T15:13:45,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741854_1030 (size=8172) 2024-12-05T15:13:45,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741854_1030 (size=8172) 2024-12-05T15:13:45,333 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/f8472c8f07634d0a95e9015d4c4ef00e, entries=3, sequenceid=48, filesize=8.0 K 2024-12-05T15:13:45,334 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~3.15 KB/3228, heapSize ~3.61 KB/3696, currentSize=0 B/0 for e2d67f93c8081f8f26afcb8999bd38fe in 33ms, sequenceid=48, compaction requested=true 2024-12-05T15:13:45,335 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.59 KB at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/.tmp/info/ac73eb4532ad40ea8f4c4c5f29e8da79 2024-12-05T15:13:45,335 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23, hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/9fb9c3efa77f47768579b2ccfc9b804c, hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/5484fedef4824de38ed42183fa1dcc97] to archive 2024-12-05T15:13:45,338 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-05T15:13:45,341 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-05T15:13:45,341 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-05T15:13:45,342 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/archive/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/7a0ab1f09fd9498e999cff4a1ca04f23 2024-12-05T15:13:45,344 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/9fb9c3efa77f47768579b2ccfc9b804c to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/archive/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/9fb9c3efa77f47768579b2ccfc9b804c 2024-12-05T15:13:45,346 DEBUG [StoreCloser-TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/5484fedef4824de38ed42183fa1dcc97 to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/archive/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/info/5484fedef4824de38ed42183fa1dcc97 2024-12-05T15:13:45,359 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/.tmp/table/b1d95481f3864fd5b95ff674c548ee54 is 73, key is TestLogRolling-testSlowSyncLogRolling/table:state/1733411518295/Put/seqid=0 2024-12-05T15:13:45,366 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741855_1031 (size=5452) 2024-12-05T15:13:45,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741855_1031 (size=5452) 2024-12-05T15:13:45,367 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=232 B at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/.tmp/table/b1d95481f3864fd5b95ff674c548ee54 2024-12-05T15:13:45,370 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/default/TestLogRolling-testSlowSyncLogRolling/e2d67f93c8081f8f26afcb8999bd38fe/recovered.edits/51.seqid, newMaxSeqId=51, maxSeqId=1 2024-12-05T15:13:45,372 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:13:45,372 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for e2d67f93c8081f8f26afcb8999bd38fe: 2024-12-05T15:13:45,373 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testSlowSyncLogRolling,,1733411517887.e2d67f93c8081f8f26afcb8999bd38fe. 2024-12-05T15:13:45,373 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 2ffa63508ac81c218f03e91221ac2e05, disabling compactions & flushes 2024-12-05T15:13:45,373 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:13:45,373 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:13:45,373 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. after waiting 0 ms 2024-12-05T15:13:45,373 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:13:45,373 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 2ffa63508ac81c218f03e91221ac2e05 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-05T15:13:45,377 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/.tmp/info/ac73eb4532ad40ea8f4c4c5f29e8da79 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/info/ac73eb4532ad40ea8f4c4c5f29e8da79 2024-12-05T15:13:45,386 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/info/ac73eb4532ad40ea8f4c4c5f29e8da79, entries=20, sequenceid=14, filesize=8.0 K 2024-12-05T15:13:45,387 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/.tmp/table/b1d95481f3864fd5b95ff674c548ee54 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/table/b1d95481f3864fd5b95ff674c548ee54 2024-12-05T15:13:45,394 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05/.tmp/info/8817dd973f2f4673935609e5a93f489d is 45, key is default/info:d/1733411517640/Put/seqid=0 2024-12-05T15:13:45,396 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/table/b1d95481f3864fd5b95ff674c548ee54, entries=4, sequenceid=14, filesize=5.3 K 2024-12-05T15:13:45,397 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~2.81 KB/2882, heapSize ~5.04 KB/5160, currentSize=0 B/0 for 1588230740 in 96ms, sequenceid=14, compaction requested=false 2024-12-05T15:13:45,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741856_1032 (size=5037) 2024-12-05T15:13:45,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741856_1032 (size=5037) 2024-12-05T15:13:45,401 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05/.tmp/info/8817dd973f2f4673935609e5a93f489d 2024-12-05T15:13:45,404 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/meta/1588230740/recovered.edits/17.seqid, newMaxSeqId=17, maxSeqId=1 2024-12-05T15:13:45,405 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T15:13:45,405 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:13:45,405 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:13:45,406 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-05T15:13:45,409 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05/.tmp/info/8817dd973f2f4673935609e5a93f489d as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05/info/8817dd973f2f4673935609e5a93f489d 2024-12-05T15:13:45,417 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05/info/8817dd973f2f4673935609e5a93f489d, entries=2, sequenceid=6, filesize=4.9 K 2024-12-05T15:13:45,418 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 2ffa63508ac81c218f03e91221ac2e05 in 45ms, sequenceid=6, compaction requested=false 2024-12-05T15:13:45,423 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/data/hbase/namespace/2ffa63508ac81c218f03e91221ac2e05/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-05T15:13:45,424 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:13:45,424 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 2ffa63508ac81c218f03e91221ac2e05: 2024-12-05T15:13:45,424 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733411517082.2ffa63508ac81c218f03e91221ac2e05. 2024-12-05T15:13:45,502 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,37867,1733411515229; all regions closed. 2024-12-05T15:13:45,503 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229 2024-12-05T15:13:45,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741834_1010 (size=4330) 2024-12-05T15:13:45,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741834_1010 (size=4330) 2024-12-05T15:13:45,510 DEBUG [RS:0;4ecb0f57b706:37867 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs 2024-12-05T15:13:45,510 INFO [RS:0;4ecb0f57b706:37867 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C37867%2C1733411515229.meta:.meta(num 1733411516845) 2024-12-05T15:13:45,510 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/WALs/4ecb0f57b706,37867,1733411515229 2024-12-05T15:13:45,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741851_1027 (size=13066) 2024-12-05T15:13:45,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741851_1027 (size=13066) 2024-12-05T15:13:45,522 DEBUG [RS:0;4ecb0f57b706:37867 {}] wal.AbstractFSWAL(1071): Moved 3 WAL file(s) to /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/oldWALs 2024-12-05T15:13:45,522 INFO [RS:0;4ecb0f57b706:37867 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C37867%2C1733411515229:(num 1733411605296) 2024-12-05T15:13:45,522 DEBUG [RS:0;4ecb0f57b706:37867 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:13:45,523 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:13:45,523 INFO [RS:0;4ecb0f57b706:37867 {}] hbase.ChoreService(370): Chore service for: regionserver/4ecb0f57b706:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-05T15:13:45,524 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:13:45,524 INFO [RS:0;4ecb0f57b706:37867 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:37867 2024-12-05T15:13:45,528 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/4ecb0f57b706,37867,1733411515229 2024-12-05T15:13:45,528 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:13:45,530 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [4ecb0f57b706,37867,1733411515229] 2024-12-05T15:13:45,530 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 4ecb0f57b706,37867,1733411515229; numProcessing=1 2024-12-05T15:13:45,531 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/4ecb0f57b706,37867,1733411515229 already deleted, retry=false 2024-12-05T15:13:45,531 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 4ecb0f57b706,37867,1733411515229 expired; onlineServers=0 2024-12-05T15:13:45,531 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,35771,1733411514503' ***** 2024-12-05T15:13:45,531 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T15:13:45,532 DEBUG [M:0;4ecb0f57b706:35771 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1d2217c1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:13:45,532 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,35771,1733411514503 2024-12-05T15:13:45,532 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,35771,1733411514503; all regions closed. 2024-12-05T15:13:45,532 DEBUG [M:0;4ecb0f57b706:35771 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:13:45,532 DEBUG [M:0;4ecb0f57b706:35771 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T15:13:45,532 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T15:13:45,532 DEBUG [M:0;4ecb0f57b706:35771 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T15:13:45,532 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411516130 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411516130,5,FailOnTimeoutGroup] 2024-12-05T15:13:45,532 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411516130 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411516130,5,FailOnTimeoutGroup] 2024-12-05T15:13:45,533 INFO [M:0;4ecb0f57b706:35771 {}] hbase.ChoreService(370): Chore service for: master/4ecb0f57b706:0 had [] on shutdown 2024-12-05T15:13:45,533 DEBUG [M:0;4ecb0f57b706:35771 {}] master.HMaster(1733): Stopping service threads 2024-12-05T15:13:45,533 INFO [M:0;4ecb0f57b706:35771 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T15:13:45,533 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T15:13:45,533 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:45,534 INFO [M:0;4ecb0f57b706:35771 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T15:13:45,534 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:13:45,534 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T15:13:45,534 DEBUG [M:0;4ecb0f57b706:35771 {}] zookeeper.ZKUtil(347): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T15:13:45,534 WARN [M:0;4ecb0f57b706:35771 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T15:13:45,534 INFO [M:0;4ecb0f57b706:35771 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-05T15:13:45,534 INFO [M:0;4ecb0f57b706:35771 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T15:13:45,534 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:13:45,535 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:45,535 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:45,535 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:13:45,535 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:45,535 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=40.21 KB heapSize=50.14 KB 2024-12-05T15:13:45,553 DEBUG [M:0;4ecb0f57b706:35771 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/87000e6cfd8242578d5691045140b6cc is 82, key is hbase:meta,,1/info:regioninfo/1733411516965/Put/seqid=0 2024-12-05T15:13:45,559 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741857_1033 (size=5672) 2024-12-05T15:13:45,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741857_1033 (size=5672) 2024-12-05T15:13:45,560 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/87000e6cfd8242578d5691045140b6cc 2024-12-05T15:13:45,589 DEBUG [M:0;4ecb0f57b706:35771 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1588af23bf0943fb9d7a25e6f5966a7b is 766, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733411518303/Put/seqid=0 2024-12-05T15:13:45,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741858_1034 (size=6426) 2024-12-05T15:13:45,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741858_1034 (size=6426) 2024-12-05T15:13:45,595 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=39.61 KB at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1588af23bf0943fb9d7a25e6f5966a7b 2024-12-05T15:13:45,601 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 1588af23bf0943fb9d7a25e6f5966a7b 2024-12-05T15:13:45,617 DEBUG [M:0;4ecb0f57b706:35771 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/5056c48dfdc04633b61f1837f75ae348 is 69, key is 4ecb0f57b706,37867,1733411515229/rs:state/1733411516269/Put/seqid=0 2024-12-05T15:13:45,623 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741859_1035 (size=5156) 2024-12-05T15:13:45,624 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741859_1035 (size=5156) 2024-12-05T15:13:45,624 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/5056c48dfdc04633b61f1837f75ae348 2024-12-05T15:13:45,630 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:13:45,630 INFO [RS:0;4ecb0f57b706:37867 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,37867,1733411515229; zookeeper connection closed. 2024-12-05T15:13:45,630 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37867-0x10062bc5e100001, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:13:45,630 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@12464540 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@12464540 2024-12-05T15:13:45,630 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-05T15:13:45,646 DEBUG [M:0;4ecb0f57b706:35771 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/ca90326bd8274301acb5761011932bc1 is 52, key is load_balancer_on/state:d/1733411517868/Put/seqid=0 2024-12-05T15:13:45,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741860_1036 (size=5056) 2024-12-05T15:13:45,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741860_1036 (size=5056) 2024-12-05T15:13:45,652 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=104 (bloomFilter=true), to=hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/ca90326bd8274301acb5761011932bc1 2024-12-05T15:13:45,658 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/87000e6cfd8242578d5691045140b6cc as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/87000e6cfd8242578d5691045140b6cc 2024-12-05T15:13:45,665 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/87000e6cfd8242578d5691045140b6cc, entries=8, sequenceid=104, filesize=5.5 K 2024-12-05T15:13:45,666 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/1588af23bf0943fb9d7a25e6f5966a7b as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/1588af23bf0943fb9d7a25e6f5966a7b 2024-12-05T15:13:45,672 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 1588af23bf0943fb9d7a25e6f5966a7b 2024-12-05T15:13:45,672 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/1588af23bf0943fb9d7a25e6f5966a7b, entries=11, sequenceid=104, filesize=6.3 K 2024-12-05T15:13:45,673 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/5056c48dfdc04633b61f1837f75ae348 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/5056c48dfdc04633b61f1837f75ae348 2024-12-05T15:13:45,680 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/5056c48dfdc04633b61f1837f75ae348, entries=1, sequenceid=104, filesize=5.0 K 2024-12-05T15:13:45,681 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/ca90326bd8274301acb5761011932bc1 as hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/ca90326bd8274301acb5761011932bc1 2024-12-05T15:13:45,692 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/ca90326bd8274301acb5761011932bc1, entries=1, sequenceid=104, filesize=4.9 K 2024-12-05T15:13:45,694 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(3040): Finished flush of dataSize ~40.21 KB/41173, heapSize ~50.08 KB/51280, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 158ms, sequenceid=104, compaction requested=false 2024-12-05T15:13:45,696 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:45,696 DEBUG [M:0;4ecb0f57b706:35771 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:13:45,696 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/MasterData/WALs/4ecb0f57b706,35771,1733411514503 2024-12-05T15:13:45,698 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46131 is added to blk_1073741830_1006 (size=48474) 2024-12-05T15:13:45,699 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40039 is added to blk_1073741830_1006 (size=48474) 2024-12-05T15:13:45,699 INFO [M:0;4ecb0f57b706:35771 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-05T15:13:45,699 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:13:45,699 INFO [M:0;4ecb0f57b706:35771 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:35771 2024-12-05T15:13:45,701 DEBUG [M:0;4ecb0f57b706:35771 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/4ecb0f57b706,35771,1733411514503 already deleted, retry=false 2024-12-05T15:13:45,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:13:45,803 INFO [M:0;4ecb0f57b706:35771 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,35771,1733411514503; zookeeper connection closed. 2024-12-05T15:13:45,803 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:35771-0x10062bc5e100000, quorum=127.0.0.1:52436, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:13:45,808 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2c482eac{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:45,810 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@408e1d66{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:13:45,810 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:13:45,810 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4debea22{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:13:45,810 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@670df016{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir/,STOPPED} 2024-12-05T15:13:45,813 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:13:45,813 WARN [BP-1798444416-172.17.0.2-1733411511367 heartbeating to localhost/127.0.0.1:35505 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:13:45,813 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:13:45,813 WARN [BP-1798444416-172.17.0.2-1733411511367 heartbeating to localhost/127.0.0.1:35505 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1798444416-172.17.0.2-1733411511367 (Datanode Uuid fb982a57-09ca-4495-ba58-d9a24c0e5bbe) service to localhost/127.0.0.1:35505 2024-12-05T15:13:45,815 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data3/current/BP-1798444416-172.17.0.2-1733411511367 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:45,815 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data4/current/BP-1798444416-172.17.0.2-1733411511367 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:45,815 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:13:45,817 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5c4b1b4{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:45,818 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@9fb4bbe{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:13:45,818 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:13:45,818 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2276bd44{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:13:45,818 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3cf7922e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir/,STOPPED} 2024-12-05T15:13:45,820 WARN [BP-1798444416-172.17.0.2-1733411511367 heartbeating to localhost/127.0.0.1:35505 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:13:45,820 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:13:45,820 WARN [BP-1798444416-172.17.0.2-1733411511367 heartbeating to localhost/127.0.0.1:35505 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1798444416-172.17.0.2-1733411511367 (Datanode Uuid 795230d8-3c86-4981-a198-c31c3ffefc5c) service to localhost/127.0.0.1:35505 2024-12-05T15:13:45,820 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:13:45,821 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data1/current/BP-1798444416-172.17.0.2-1733411511367 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:45,821 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/cluster_92a8192b-cd44-3617-b654-e7d0874be9ef/dfs/data/data2/current/BP-1798444416-172.17.0.2-1733411511367 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:45,821 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:13:45,830 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5682c4d1{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:13:45,831 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1ff1a6c1{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:13:45,831 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:13:45,831 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@74468826{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:13:45,831 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@88aab13{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir/,STOPPED} 2024-12-05T15:13:45,841 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-05T15:13:45,894 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-05T15:13:45,904 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testSlowSyncLogRolling Thread=65 (was 12) Potentially hanging thread: Timer for 'HBase' metrics system java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: Time-limited test.named-queue-events-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:35505 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-4-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: master/4ecb0f57b706:0:becomeActiveMaster-MemStoreChunkPool Statistics java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-2-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Potentially hanging thread: SessionTracker java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Potentially hanging thread: nioEventLoopGroup-4-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-5-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-3-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: master/4ecb0f57b706:0:becomeActiveMaster-MemStoreChunkPool Statistics java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: SnapshotHandlerChoreCleaner java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.0@localhost:35505 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:35505 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: Async-Client-Retry-Timer-pool-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-2-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-4-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: region-location-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: regionserver/4ecb0f57b706:0.procedureResultReporter java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Potentially hanging thread: region-location-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: ForkJoinPool-2-worker-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: LeaseRenewer:jenkins@localhost:35505 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-5-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:35505 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: Idle-Rpc-Conn-Sweeper-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: SSL Certificates Store Monitor java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: RpcClient-timer-pool-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: org.apache.hadoop.hdfs.PeerCache@1a9fc923 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:35505 from jenkins.hfs.0 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: Monitor thread for TaskMonitor java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-2 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-3-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-1-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: regionserver/4ecb0f57b706:0.leaseChecker java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.regionserver.LeaseManager.run(LeaseManager.java:82) Potentially hanging thread: RS-EventLoopGroup-1-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:35505 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:35505 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: RS-EventLoopGroup-3-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:35505 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: nioEventLoopGroup-3-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-1-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HBase-Metrics2-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool-2-worker-3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-1 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-2-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-5-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=401 (was 286) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=226 (was 363), ProcessCount=11 (was 11), AvailableMemoryMB=9072 (was 9665) 2024-12-05T15:13:45,914 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnDatanodeDeath Thread=66, OpenFileDescriptor=401, MaxFileDescriptor=1048576, SystemLoadAverage=226, ProcessCount=11, AvailableMemoryMB=9072 2024-12-05T15:13:45,914 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T15:13:45,914 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.log.dir so I do NOT create it in target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f 2024-12-05T15:13:45,914 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/1b983e05-e281-2c87-226a-72e8fc926c86/hadoop.tmp.dir so I do NOT create it in target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f 2024-12-05T15:13:45,914 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95, deleteOnExit=true 2024-12-05T15:13:45,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-05T15:13:45,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/test.cache.data in system properties and HBase conf 2024-12-05T15:13:45,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T15:13:45,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir in system properties and HBase conf 2024-12-05T15:13:45,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T15:13:45,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T15:13:45,915 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-05T15:13:45,915 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T15:13:45,916 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:13:45,916 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:13:45,916 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T15:13:45,916 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:13:45,916 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T15:13:45,917 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T15:13:45,917 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:13:45,917 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:13:45,917 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T15:13:45,917 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/nfs.dump.dir in system properties and HBase conf 2024-12-05T15:13:45,917 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir in system properties and HBase conf 2024-12-05T15:13:45,917 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:13:45,918 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T15:13:45,918 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T15:13:45,935 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:13:46,022 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:13:46,028 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:13:46,029 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:13:46,029 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:13:46,029 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:13:46,030 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:13:46,031 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1101772a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:13:46,031 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2c0370f2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:13:46,150 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@55bfb256{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir/jetty-localhost-42161-hadoop-hdfs-3_4_1-tests_jar-_-any-8124647134922108043/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:13:46,150 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@7e1ef301{HTTP/1.1, (http/1.1)}{localhost:42161} 2024-12-05T15:13:46,151 INFO [Time-limited test {}] server.Server(415): Started @116966ms 2024-12-05T15:13:46,166 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:13:46,247 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:13:46,250 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:13:46,251 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:13:46,251 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:13:46,251 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:13:46,252 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@59774852{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:13:46,252 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@220b71ae{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:13:46,345 INFO [regionserver/4ecb0f57b706:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:13:46,377 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5ecf65b2{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir/jetty-localhost-39399-hadoop-hdfs-3_4_1-tests_jar-_-any-11003643444603341800/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:46,378 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5483975{HTTP/1.1, (http/1.1)}{localhost:39399} 2024-12-05T15:13:46,378 INFO [Time-limited test {}] server.Server(415): Started @117194ms 2024-12-05T15:13:46,380 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:13:46,419 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:13:46,422 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:13:46,423 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:13:46,423 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:13:46,423 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:13:46,424 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@29f7fced{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:13:46,424 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@60b15398{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:13:46,479 WARN [Thread-450 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data1/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:46,479 WARN [Thread-451 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data2/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:46,503 WARN [Thread-429 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:13:46,507 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc2d61ce097888f73 with lease ID 0xeb533680bbc0558b: Processing first storage report for DS-5459db68-20a8-4270-b89c-e411daaa5d88 from datanode DatanodeRegistration(127.0.0.1:43797, datanodeUuid=827e5892-6c93-4960-ae7d-59bc773f1ed4, infoPort=39611, infoSecurePort=0, ipcPort=44937, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:46,507 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc2d61ce097888f73 with lease ID 0xeb533680bbc0558b: from storage DS-5459db68-20a8-4270-b89c-e411daaa5d88 node DatanodeRegistration(127.0.0.1:43797, datanodeUuid=827e5892-6c93-4960-ae7d-59bc773f1ed4, infoPort=39611, infoSecurePort=0, ipcPort=44937, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-05T15:13:46,507 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xc2d61ce097888f73 with lease ID 0xeb533680bbc0558b: Processing first storage report for DS-df6af8fa-653e-4088-8dd1-52f988f7535a from datanode DatanodeRegistration(127.0.0.1:43797, datanodeUuid=827e5892-6c93-4960-ae7d-59bc773f1ed4, infoPort=39611, infoSecurePort=0, ipcPort=44937, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:46,507 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xc2d61ce097888f73 with lease ID 0xeb533680bbc0558b: from storage DS-df6af8fa-653e-4088-8dd1-52f988f7535a node DatanodeRegistration(127.0.0.1:43797, datanodeUuid=827e5892-6c93-4960-ae7d-59bc773f1ed4, infoPort=39611, infoSecurePort=0, ipcPort=44937, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:46,547 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@2d70ba54{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir/jetty-localhost-45509-hadoop-hdfs-3_4_1-tests_jar-_-any-1136825163645134079/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:46,547 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1a6c4f80{HTTP/1.1, (http/1.1)}{localhost:45509} 2024-12-05T15:13:46,547 INFO [Time-limited test {}] server.Server(415): Started @117363ms 2024-12-05T15:13:46,549 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:13:46,656 WARN [Thread-476 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data3/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:46,657 WARN [Thread-477 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data4/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:46,680 WARN [Thread-465 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:13:46,683 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb447a4f07e8a2b08 with lease ID 0xeb533680bbc0558c: Processing first storage report for DS-6781d264-5b19-4912-a89b-9a138120ec6a from datanode DatanodeRegistration(127.0.0.1:41001, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=34537, infoSecurePort=0, ipcPort=34367, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:46,683 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb447a4f07e8a2b08 with lease ID 0xeb533680bbc0558c: from storage DS-6781d264-5b19-4912-a89b-9a138120ec6a node DatanodeRegistration(127.0.0.1:41001, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=34537, infoSecurePort=0, ipcPort=34367, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:46,684 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb447a4f07e8a2b08 with lease ID 0xeb533680bbc0558c: Processing first storage report for DS-39ac30c5-c7cf-427d-8a5c-40fc28623f09 from datanode DatanodeRegistration(127.0.0.1:41001, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=34537, infoSecurePort=0, ipcPort=34367, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:46,684 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb447a4f07e8a2b08 with lease ID 0xeb533680bbc0558c: from storage DS-39ac30c5-c7cf-427d-8a5c-40fc28623f09 node DatanodeRegistration(127.0.0.1:41001, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=34537, infoSecurePort=0, ipcPort=34367, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:46,782 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f 2024-12-05T15:13:46,785 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/zookeeper_0, clientPort=55344, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T15:13:46,787 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=55344 2024-12-05T15:13:46,787 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:46,789 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:46,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:13:46,801 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:13:46,802 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b with version=8 2024-12-05T15:13:46,802 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/hbase-staging 2024-12-05T15:13:46,804 INFO [Time-limited test {}] client.ConnectionUtils(129): master/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:13:46,804 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:46,804 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:46,804 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:13:46,805 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:46,805 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:13:46,805 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:13:46,805 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:13:46,806 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:38149 2024-12-05T15:13:46,806 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:46,808 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:46,811 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:38149 connecting to ZooKeeper ensemble=127.0.0.1:55344 2024-12-05T15:13:46,817 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:381490x0, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:13:46,818 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:38149-0x10062be18070000 connected 2024-12-05T15:13:46,838 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:13:46,839 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:13:46,840 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:13:46,840 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=38149 2024-12-05T15:13:46,840 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=38149 2024-12-05T15:13:46,840 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=38149 2024-12-05T15:13:46,841 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=38149 2024-12-05T15:13:46,841 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=38149 2024-12-05T15:13:46,841 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b, hbase.cluster.distributed=false 2024-12-05T15:13:46,859 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:13:46,860 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:46,860 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:46,860 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:13:46,860 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:46,860 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:13:46,860 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:13:46,860 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:13:46,861 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:44787 2024-12-05T15:13:46,862 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T15:13:46,862 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T15:13:46,863 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:46,866 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:46,869 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:44787 connecting to ZooKeeper ensemble=127.0.0.1:55344 2024-12-05T15:13:46,871 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:447870x0, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:13:46,872 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:44787-0x10062be18070001 connected 2024-12-05T15:13:46,872 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:13:46,873 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:13:46,873 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:13:46,876 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=44787 2024-12-05T15:13:46,877 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=44787 2024-12-05T15:13:46,879 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=44787 2024-12-05T15:13:46,879 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=44787 2024-12-05T15:13:46,880 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=44787 2024-12-05T15:13:46,882 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:46,884 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:13:46,884 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:13:46,884 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:46,886 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:13:46,886 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:46,886 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:13:46,886 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:46,887 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:13:46,887 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/4ecb0f57b706,38149,1733411626804 from backup master directory 2024-12-05T15:13:46,889 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:46,889 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:13:46,889 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:13:46,889 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:13:46,889 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:13:46,889 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:46,898 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;4ecb0f57b706:38149 2024-12-05T15:13:46,907 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:13:46,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:13:46,909 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/hbase.id with ID: daa5ecc0-4acd-4ba9-9bb7-ca16eb42614b 2024-12-05T15:13:46,924 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:46,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:46,929 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:46,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:13:46,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:13:46,945 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:13:46,946 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T15:13:46,947 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:13:46,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:13:46,959 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:13:46,961 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store 2024-12-05T15:13:46,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:13:46,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:13:46,972 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:46,972 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:13:46,972 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:46,972 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:46,972 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:13:46,972 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:46,972 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:13:46,972 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:13:46,973 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/.initializing 2024-12-05T15:13:46,973 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:46,978 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C38149%2C1733411626804, suffix=, logDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804, archiveDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/oldWALs, maxLogs=10 2024-12-05T15:13:46,978 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C38149%2C1733411626804.1733411626978 2024-12-05T15:13:46,989 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 2024-12-05T15:13:46,989 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34537:34537),(127.0.0.1/127.0.0.1:39611:39611)] 2024-12-05T15:13:46,989 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:13:46,989 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:46,989 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:46,989 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:46,991 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:46,993 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T15:13:46,993 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:46,994 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:13:46,994 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:46,995 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T15:13:46,996 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:46,996 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:13:46,996 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:46,998 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T15:13:46,998 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:46,998 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:13:46,998 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:47,001 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T15:13:47,001 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,001 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:13:47,002 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:47,003 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:47,005 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T15:13:47,006 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:13:47,009 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:13:47,009 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=788472, jitterRate=0.0025947242975234985}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T15:13:47,011 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:13:47,011 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T15:13:47,015 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@79d69d2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:13:47,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-05T15:13:47,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T15:13:47,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T15:13:47,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T15:13:47,018 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 1 msec 2024-12-05T15:13:47,018 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-05T15:13:47,018 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T15:13:47,021 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T15:13:47,022 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T15:13:47,024 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-05T15:13:47,024 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T15:13:47,025 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T15:13:47,026 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-05T15:13:47,026 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T15:13:47,027 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T15:13:47,029 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-05T15:13:47,029 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T15:13:47,031 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T15:13:47,033 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T15:13:47,034 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T15:13:47,036 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:13:47,036 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:13:47,036 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,036 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,037 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=4ecb0f57b706,38149,1733411626804, sessionid=0x10062be18070000, setting cluster-up flag (Was=false) 2024-12-05T15:13:47,040 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,040 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,046 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T15:13:47,047 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:47,051 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,051 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,056 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T15:13:47,058 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:47,062 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-05T15:13:47,062 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-05T15:13:47,062 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 4ecb0f57b706,38149,1733411626804 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/4ecb0f57b706:0, corePoolSize=10, maxPoolSize=10 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:13:47,064 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,066 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733411657066 2024-12-05T15:13:47,066 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T15:13:47,066 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T15:13:47,066 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:13:47,066 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T15:13:47,066 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T15:13:47,066 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T15:13:47,066 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-05T15:13:47,066 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T15:13:47,067 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,067 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T15:13:47,067 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T15:13:47,067 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T15:13:47,068 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T15:13:47,068 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T15:13:47,068 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411627068,5,FailOnTimeoutGroup] 2024-12-05T15:13:47,068 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,069 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411627068,5,FailOnTimeoutGroup] 2024-12-05T15:13:47,069 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,069 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T15:13:47,069 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,069 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:13:47,069 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:13:47,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:13:47,085 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-05T15:13:47,086 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b 2024-12-05T15:13:47,099 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;4ecb0f57b706:44787 2024-12-05T15:13:47,101 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1008): ClusterId : daa5ecc0-4acd-4ba9-9bb7-ca16eb42614b 2024-12-05T15:13:47,101 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T15:13:47,105 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T15:13:47,105 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T15:13:47,106 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:13:47,106 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:13:47,108 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:47,110 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:13:47,111 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T15:13:47,111 DEBUG [RS:0;4ecb0f57b706:44787 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@65a51377, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:13:47,112 DEBUG [RS:0;4ecb0f57b706:44787 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2605b858, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:13:47,112 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-05T15:13:47,112 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-05T15:13:47,112 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-05T15:13:47,115 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,38149,1733411626804 with isa=4ecb0f57b706/172.17.0.2:44787, startcode=1733411626859 2024-12-05T15:13:47,115 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:13:47,115 DEBUG [RS:0;4ecb0f57b706:44787 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:13:47,115 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,117 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:13:47,117 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:13:47,119 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:13:47,119 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,121 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44991, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:13:47,121 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:13:47,121 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:13:47,122 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38149 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,122 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38149 {}] master.ServerManager(486): Registering regionserver=4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,125 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:13:47,125 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,125 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b 2024-12-05T15:13:47,125 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:13:47,127 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/meta/1588230740 2024-12-05T15:13:47,127 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/meta/1588230740 2024-12-05T15:13:47,125 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:32987 2024-12-05T15:13:47,128 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-05T15:13:47,130 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:13:47,131 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:13:47,131 DEBUG [RS:0;4ecb0f57b706:44787 {}] zookeeper.ZKUtil(111): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,131 WARN [RS:0;4ecb0f57b706:44787 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:13:47,132 INFO [RS:0;4ecb0f57b706:44787 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:13:47,132 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,132 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [4ecb0f57b706,44787,1733411626859] 2024-12-05T15:13:47,133 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:13:47,136 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-05T15:13:47,136 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T15:13:47,137 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:13:47,138 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=752003, jitterRate=-0.04377928376197815}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:13:47,140 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:13:47,140 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:13:47,140 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:13:47,140 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:13:47,140 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:13:47,140 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:13:47,142 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T15:13:47,142 INFO [RS:0;4ecb0f57b706:44787 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:13:47,142 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,144 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-05T15:13:47,145 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,145 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,146 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,146 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,146 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,146 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,146 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:13:47,146 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,147 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,147 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,147 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,147 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:47,147 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:13:47,147 DEBUG [RS:0;4ecb0f57b706:44787 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:13:47,148 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:13:47,148 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:13:47,148 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,149 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,149 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,149 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,149 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,44787,1733411626859-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:13:47,151 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:13:47,151 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-05T15:13:47,151 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T15:13:47,153 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T15:13:47,155 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T15:13:47,168 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T15:13:47,169 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,44787,1733411626859-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,184 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.Replication(204): 4ecb0f57b706,44787,1733411626859 started 2024-12-05T15:13:47,184 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1767): Serving as 4ecb0f57b706,44787,1733411626859, RpcServer on 4ecb0f57b706/172.17.0.2:44787, sessionid=0x10062be18070001 2024-12-05T15:13:47,184 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T15:13:47,184 DEBUG [RS:0;4ecb0f57b706:44787 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,185 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,44787,1733411626859' 2024-12-05T15:13:47,185 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T15:13:47,185 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T15:13:47,186 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T15:13:47,186 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T15:13:47,186 DEBUG [RS:0;4ecb0f57b706:44787 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,186 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,44787,1733411626859' 2024-12-05T15:13:47,186 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T15:13:47,186 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T15:13:47,187 DEBUG [RS:0;4ecb0f57b706:44787 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T15:13:47,187 INFO [RS:0;4ecb0f57b706:44787 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T15:13:47,187 INFO [RS:0;4ecb0f57b706:44787 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T15:13:47,290 INFO [RS:0;4ecb0f57b706:44787 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C44787%2C1733411626859, suffix=, logDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859, archiveDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs, maxLogs=32 2024-12-05T15:13:47,294 INFO [RS:0;4ecb0f57b706:44787 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.1733411627293 2024-12-05T15:13:47,306 WARN [4ecb0f57b706:38149 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-05T15:13:47,324 INFO [RS:0;4ecb0f57b706:44787 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 2024-12-05T15:13:47,324 DEBUG [RS:0;4ecb0f57b706:44787 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34537:34537),(127.0.0.1/127.0.0.1:39611:39611)] 2024-12-05T15:13:47,556 DEBUG [4ecb0f57b706:38149 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-05T15:13:47,556 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,558 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,44787,1733411626859, state=OPENING 2024-12-05T15:13:47,559 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T15:13:47,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,561 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:47,562 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=4ecb0f57b706,44787,1733411626859}] 2024-12-05T15:13:47,562 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:13:47,562 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:13:47,714 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,714 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T15:13:47,717 INFO [RS-EventLoopGroup-6-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:57746, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T15:13:47,722 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-05T15:13:47,722 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:13:47,724 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C44787%2C1733411626859.meta, suffix=.meta, logDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859, archiveDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs, maxLogs=32 2024-12-05T15:13:47,726 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta 2024-12-05T15:13:47,738 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta 2024-12-05T15:13:47,738 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:34537:34537),(127.0.0.1/127.0.0.1:39611:39611)] 2024-12-05T15:13:47,738 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:13:47,738 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T15:13:47,738 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T15:13:47,739 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T15:13:47,739 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T15:13:47,739 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:47,739 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-05T15:13:47,739 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-05T15:13:47,741 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:13:47,742 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:13:47,742 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,742 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:13:47,743 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:13:47,744 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:13:47,744 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,744 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:13:47,744 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:13:47,745 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:13:47,745 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,746 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:13:47,747 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/meta/1588230740 2024-12-05T15:13:47,748 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/meta/1588230740 2024-12-05T15:13:47,750 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:13:47,751 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:13:47,753 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=783513, jitterRate=-0.003712683916091919}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:13:47,754 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:13:47,755 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733411627714 2024-12-05T15:13:47,758 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T15:13:47,758 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-05T15:13:47,759 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:47,760 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,44787,1733411626859, state=OPEN 2024-12-05T15:13:47,765 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:13:47,765 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:13:47,765 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:13:47,765 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:13:47,768 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T15:13:47,768 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=4ecb0f57b706,44787,1733411626859 in 203 msec 2024-12-05T15:13:47,772 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T15:13:47,772 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 617 msec 2024-12-05T15:13:47,775 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 713 msec 2024-12-05T15:13:47,775 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733411627775, completionTime=-1 2024-12-05T15:13:47,775 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-05T15:13:47,775 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-05T15:13:47,776 DEBUG [hconnection-0x244786cc-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:13:47,777 INFO [RS-EventLoopGroup-6-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:57756, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:13:47,778 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-05T15:13:47,779 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733411687779 2024-12-05T15:13:47,779 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733411747779 2024-12-05T15:13:47,779 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 3 msec 2024-12-05T15:13:47,787 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,38149,1733411626804-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,787 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,38149,1733411626804-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,787 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,38149,1733411626804-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,787 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-4ecb0f57b706:38149, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,787 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:47,787 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-05T15:13:47,787 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:13:47,795 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-05T15:13:47,795 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-05T15:13:47,797 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:13:47,797 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:47,798 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:13:47,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:13:47,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:13:47,820 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 474a835416f2f984a3141295bfac95bd, NAME => 'hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b 2024-12-05T15:13:47,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:13:47,847 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:13:47,847 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:47,847 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 474a835416f2f984a3141295bfac95bd, disabling compactions & flushes 2024-12-05T15:13:47,847 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:13:47,847 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:13:47,847 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. after waiting 0 ms 2024-12-05T15:13:47,847 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:13:47,847 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:13:47,847 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 474a835416f2f984a3141295bfac95bd: 2024-12-05T15:13:47,849 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:13:47,849 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733411627849"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411627849"}]},"ts":"1733411627849"} 2024-12-05T15:13:47,852 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:13:47,853 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:13:47,854 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411627853"}]},"ts":"1733411627853"} 2024-12-05T15:13:47,855 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-05T15:13:47,859 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=474a835416f2f984a3141295bfac95bd, ASSIGN}] 2024-12-05T15:13:47,861 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=474a835416f2f984a3141295bfac95bd, ASSIGN 2024-12-05T15:13:47,862 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=474a835416f2f984a3141295bfac95bd, ASSIGN; state=OFFLINE, location=4ecb0f57b706,44787,1733411626859; forceNewPlan=false, retain=false 2024-12-05T15:13:48,013 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=474a835416f2f984a3141295bfac95bd, regionState=OPENING, regionLocation=4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:48,016 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 474a835416f2f984a3141295bfac95bd, server=4ecb0f57b706,44787,1733411626859}] 2024-12-05T15:13:48,169 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:48,174 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:13:48,174 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 474a835416f2f984a3141295bfac95bd, NAME => 'hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:13:48,175 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 474a835416f2f984a3141295bfac95bd 2024-12-05T15:13:48,175 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:48,175 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 474a835416f2f984a3141295bfac95bd 2024-12-05T15:13:48,175 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 474a835416f2f984a3141295bfac95bd 2024-12-05T15:13:48,176 INFO [StoreOpener-474a835416f2f984a3141295bfac95bd-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 474a835416f2f984a3141295bfac95bd 2024-12-05T15:13:48,178 INFO [StoreOpener-474a835416f2f984a3141295bfac95bd-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 474a835416f2f984a3141295bfac95bd columnFamilyName info 2024-12-05T15:13:48,178 DEBUG [StoreOpener-474a835416f2f984a3141295bfac95bd-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:48,179 INFO [StoreOpener-474a835416f2f984a3141295bfac95bd-1 {}] regionserver.HStore(327): Store=474a835416f2f984a3141295bfac95bd/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:13:48,180 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd 2024-12-05T15:13:48,180 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd 2024-12-05T15:13:48,182 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 474a835416f2f984a3141295bfac95bd 2024-12-05T15:13:48,185 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:13:48,186 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 474a835416f2f984a3141295bfac95bd; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=748360, jitterRate=-0.04841151833534241}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:13:48,187 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 474a835416f2f984a3141295bfac95bd: 2024-12-05T15:13:48,188 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd., pid=6, masterSystemTime=1733411628169 2024-12-05T15:13:48,190 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:13:48,190 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:13:48,191 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=474a835416f2f984a3141295bfac95bd, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:48,196 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T15:13:48,196 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 474a835416f2f984a3141295bfac95bd, server=4ecb0f57b706,44787,1733411626859 in 177 msec 2024-12-05T15:13:48,199 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T15:13:48,199 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=474a835416f2f984a3141295bfac95bd, ASSIGN in 337 msec 2024-12-05T15:13:48,200 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:13:48,200 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411628200"}]},"ts":"1733411628200"} 2024-12-05T15:13:48,202 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-05T15:13:48,205 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:13:48,207 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 417 msec 2024-12-05T15:13:48,297 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-05T15:13:48,299 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:48,299 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:13:48,299 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:13:48,305 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-05T15:13:48,324 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:13:48,329 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 24 msec 2024-12-05T15:13:48,338 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-05T15:13:48,350 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:13:48,357 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 18 msec 2024-12-05T15:13:48,365 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-05T15:13:48,368 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-05T15:13:48,368 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.479sec 2024-12-05T15:13:48,368 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T15:13:48,368 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T15:13:48,368 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T15:13:48,368 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T15:13:48,368 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T15:13:48,368 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,38149,1733411626804-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:13:48,369 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,38149,1733411626804-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T15:13:48,371 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-05T15:13:48,371 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T15:13:48,371 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,38149,1733411626804-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,385 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6fb20afc to 127.0.0.1:55344 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@67463d7c 2024-12-05T15:13:48,390 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4a43ea9a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:13:48,392 DEBUG [hconnection-0x6413a7fa-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:13:48,394 INFO [RS-EventLoopGroup-6-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:57764, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:13:48,396 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=4ecb0f57b706,38149,1733411626804 2024-12-05T15:13:48,397 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:48,400 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-05T15:13:48,417 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:13:48,417 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:48,417 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:48,417 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:13:48,417 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:13:48,417 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:13:48,417 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:13:48,418 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:13:48,418 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:39169 2024-12-05T15:13:48,419 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T15:13:48,419 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T15:13:48,420 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:48,422 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:13:48,425 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:39169 connecting to ZooKeeper ensemble=127.0.0.1:55344 2024-12-05T15:13:48,428 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:391690x0, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:13:48,428 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39169-0x10062be18070003 connected 2024-12-05T15:13:48,428 DEBUG [Time-limited test {}] zookeeper.ZKUtil(111): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:13:48,429 DEBUG [Time-limited test {}] zookeeper.ZKUtil(111): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on existing znode=/hbase/running 2024-12-05T15:13:48,430 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:13:48,431 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39169 2024-12-05T15:13:48,431 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39169 2024-12-05T15:13:48,432 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39169 2024-12-05T15:13:48,434 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39169 2024-12-05T15:13:48,434 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39169 2024-12-05T15:13:48,435 DEBUG [pool-282-thread-1 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: INIT 2024-12-05T15:13:48,447 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;4ecb0f57b706:39169 2024-12-05T15:13:48,448 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1008): ClusterId : daa5ecc0-4acd-4ba9-9bb7-ca16eb42614b 2024-12-05T15:13:48,448 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T15:13:48,450 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T15:13:48,450 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T15:13:48,452 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T15:13:48,452 DEBUG [RS:1;4ecb0f57b706:39169 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@84cb4f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:13:48,453 DEBUG [RS:1;4ecb0f57b706:39169 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2fff8901, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:13:48,453 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-05T15:13:48,453 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-05T15:13:48,453 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-05T15:13:48,454 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,38149,1733411626804 with isa=4ecb0f57b706/172.17.0.2:39169, startcode=1733411628416 2024-12-05T15:13:48,454 DEBUG [RS:1;4ecb0f57b706:39169 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:13:48,456 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:40239, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:13:48,457 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38149 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 4ecb0f57b706,39169,1733411628416 2024-12-05T15:13:48,457 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=38149 {}] master.ServerManager(486): Registering regionserver=4ecb0f57b706,39169,1733411628416 2024-12-05T15:13:48,458 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b 2024-12-05T15:13:48,458 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:32987 2024-12-05T15:13:48,459 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-05T15:13:48,460 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:13:48,461 DEBUG [RS:1;4ecb0f57b706:39169 {}] zookeeper.ZKUtil(111): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/4ecb0f57b706,39169,1733411628416 2024-12-05T15:13:48,461 WARN [RS:1;4ecb0f57b706:39169 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:13:48,461 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [4ecb0f57b706,39169,1733411628416] 2024-12-05T15:13:48,461 INFO [RS:1;4ecb0f57b706:39169 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:13:48,461 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,39169,1733411628416 2024-12-05T15:13:48,467 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-05T15:13:48,467 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T15:13:48,470 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T15:13:48,471 INFO [RS:1;4ecb0f57b706:39169 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:13:48,471 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,474 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-05T15:13:48,475 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,475 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,475 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,475 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,475 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,475 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:13:48,476 DEBUG [RS:1;4ecb0f57b706:39169 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:13:48,478 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,478 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,478 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,478 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,478 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39169,1733411628416-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:13:48,503 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T15:13:48,503 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39169,1733411628416-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:13:48,520 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.Replication(204): 4ecb0f57b706,39169,1733411628416 started 2024-12-05T15:13:48,520 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1767): Serving as 4ecb0f57b706,39169,1733411628416, RpcServer on 4ecb0f57b706/172.17.0.2:39169, sessionid=0x10062be18070003 2024-12-05T15:13:48,520 INFO [Time-limited test {}] hbase.HBaseTestingUtility(3355): Started new server=Thread[RS:1;4ecb0f57b706:39169,5,FailOnTimeoutGroup] 2024-12-05T15:13:48,520 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T15:13:48,520 DEBUG [RS:1;4ecb0f57b706:39169 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 4ecb0f57b706,39169,1733411628416 2024-12-05T15:13:48,520 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,39169,1733411628416' 2024-12-05T15:13:48,520 INFO [Time-limited test {}] wal.TestLogRolling(191): Replication=2 2024-12-05T15:13:48,520 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T15:13:48,521 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T15:13:48,522 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T15:13:48,522 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-05T15:13:48,522 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T15:13:48,522 DEBUG [RS:1;4ecb0f57b706:39169 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 4ecb0f57b706,39169,1733411628416 2024-12-05T15:13:48,522 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,39169,1733411628416' 2024-12-05T15:13:48,522 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T15:13:48,522 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T15:13:48,522 DEBUG [RS:1;4ecb0f57b706:39169 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T15:13:48,523 INFO [RS:1;4ecb0f57b706:39169 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T15:13:48,523 INFO [RS:1;4ecb0f57b706:39169 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T15:13:48,526 INFO [RS-EventLoopGroup-5-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:47412, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-05T15:13:48,527 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38149 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-05T15:13:48,527 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38149 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-05T15:13:48,528 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38149 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testLogRollOnDatanodeDeath', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:13:48,529 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38149 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath 2024-12-05T15:13:48,531 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:13:48,531 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:48,531 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38149 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRollOnDatanodeDeath" procId is: 9 2024-12-05T15:13:48,532 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38149 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:13:48,532 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:13:48,542 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741837_1013 (size=393) 2024-12-05T15:13:48,542 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741837_1013 (size=393) 2024-12-05T15:13:48,544 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 817a20dfbf0d1531445f95325bcaf27c, NAME => 'TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRollOnDatanodeDeath', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b 2024-12-05T15:13:48,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41001 is added to blk_1073741838_1014 (size=76) 2024-12-05T15:13:48,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43797 is added to blk_1073741838_1014 (size=76) 2024-12-05T15:13:48,552 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:48,552 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1681): Closing 817a20dfbf0d1531445f95325bcaf27c, disabling compactions & flushes 2024-12-05T15:13:48,553 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:48,553 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:48,553 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. after waiting 0 ms 2024-12-05T15:13:48,553 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:48,553 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:48,553 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnDatanodeDeath-pool-0 {}] regionserver.HRegion(1635): Region close journal for 817a20dfbf0d1531445f95325bcaf27c: 2024-12-05T15:13:48,554 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:13:48,554 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c.","families":{"info":[{"qualifier":"regioninfo","vlen":75,"tag":[],"timestamp":"1733411628554"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411628554"}]},"ts":"1733411628554"} 2024-12-05T15:13:48,556 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:13:48,557 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:13:48,558 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnDatanodeDeath","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411628558"}]},"ts":"1733411628558"} 2024-12-05T15:13:48,559 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnDatanodeDeath, state=ENABLING in hbase:meta 2024-12-05T15:13:48,564 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=817a20dfbf0d1531445f95325bcaf27c, ASSIGN}] 2024-12-05T15:13:48,566 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=817a20dfbf0d1531445f95325bcaf27c, ASSIGN 2024-12-05T15:13:48,567 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=817a20dfbf0d1531445f95325bcaf27c, ASSIGN; state=OFFLINE, location=4ecb0f57b706,44787,1733411626859; forceNewPlan=false, retain=false 2024-12-05T15:13:48,625 INFO [RS:1;4ecb0f57b706:39169 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C39169%2C1733411628416, suffix=, logDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,39169,1733411628416, archiveDir=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs, maxLogs=32 2024-12-05T15:13:48,627 INFO [RS:1;4ecb0f57b706:39169 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C39169%2C1733411628416.1733411628627 2024-12-05T15:13:48,635 INFO [RS:1;4ecb0f57b706:39169 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,39169,1733411628416/4ecb0f57b706%2C39169%2C1733411628416.1733411628627 2024-12-05T15:13:48,635 DEBUG [RS:1;4ecb0f57b706:39169 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39611:39611),(127.0.0.1/127.0.0.1:34537:34537)] 2024-12-05T15:13:48,718 INFO [4ecb0f57b706:38149 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2024-12-05T15:13:48,719 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=817a20dfbf0d1531445f95325bcaf27c, regionState=OPENING, regionLocation=4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:48,722 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 817a20dfbf0d1531445f95325bcaf27c, server=4ecb0f57b706,44787,1733411626859}] 2024-12-05T15:13:48,875 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:48,880 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:48,880 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 817a20dfbf0d1531445f95325bcaf27c, NAME => 'TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:13:48,881 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRollOnDatanodeDeath 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:13:48,881 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:13:48,882 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:13:48,882 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:13:48,884 INFO [StoreOpener-817a20dfbf0d1531445f95325bcaf27c-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:13:48,886 INFO [StoreOpener-817a20dfbf0d1531445f95325bcaf27c-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 817a20dfbf0d1531445f95325bcaf27c columnFamilyName info 2024-12-05T15:13:48,886 DEBUG [StoreOpener-817a20dfbf0d1531445f95325bcaf27c-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:13:48,886 INFO [StoreOpener-817a20dfbf0d1531445f95325bcaf27c-1 {}] regionserver.HStore(327): Store=817a20dfbf0d1531445f95325bcaf27c/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:13:48,887 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:13:48,888 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:13:48,890 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:13:48,892 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:13:48,893 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened 817a20dfbf0d1531445f95325bcaf27c; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=831243, jitterRate=0.0569809228181839}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:13:48,894 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 817a20dfbf0d1531445f95325bcaf27c: 2024-12-05T15:13:48,895 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c., pid=11, masterSystemTime=1733411628875 2024-12-05T15:13:48,898 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:48,898 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:48,899 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=817a20dfbf0d1531445f95325bcaf27c, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,44787,1733411626859 2024-12-05T15:13:48,905 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-05T15:13:48,905 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 817a20dfbf0d1531445f95325bcaf27c, server=4ecb0f57b706,44787,1733411626859 in 180 msec 2024-12-05T15:13:48,908 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-05T15:13:48,908 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnDatanodeDeath, region=817a20dfbf0d1531445f95325bcaf27c, ASSIGN in 341 msec 2024-12-05T15:13:48,909 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:13:48,910 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnDatanodeDeath","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411628909"}]},"ts":"1733411628909"} 2024-12-05T15:13:48,912 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnDatanodeDeath, state=ENABLED in hbase:meta 2024-12-05T15:13:48,915 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:13:48,917 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testLogRollOnDatanodeDeath in 387 msec 2024-12-05T15:13:50,374 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:13:50,382 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:13:50,895 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:13:50,897 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:13:50,920 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:13:53,136 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-05T15:13:53,137 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-05T15:13:53,138 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRollOnDatanodeDeath' 2024-12-05T15:13:54,991 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnDatanodeDeath 2024-12-05T15:13:54,991 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnDatanodeDeath Metrics about Tables on a single HBase RegionServer 2024-12-05T15:13:54,992 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testSlowSyncLogRolling 2024-12-05T15:13:58,534 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=38149 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:13:58,534 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testLogRollOnDatanodeDeath, procId: 9 completed 2024-12-05T15:13:58,538 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testLogRollOnDatanodeDeath 2024-12-05T15:13:58,538 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:13:58,552 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:13:58,555 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:13:58,556 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:13:58,556 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:13:58,557 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:13:58,557 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3a5b3132{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:13:58,558 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6e33bbd1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:13:58,682 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@75f9f533{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir/jetty-localhost-37199-hadoop-hdfs-3_4_1-tests_jar-_-any-5183092016118107659/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:58,683 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@154ff60c{HTTP/1.1, (http/1.1)}{localhost:37199} 2024-12-05T15:13:58,683 INFO [Time-limited test {}] server.Server(415): Started @129498ms 2024-12-05T15:13:58,685 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:13:58,729 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:13:58,734 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:13:58,735 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:13:58,735 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:13:58,735 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:13:58,736 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@493e02d8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:13:58,737 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@55167cef{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:13:58,780 WARN [Thread-633 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:58,780 WARN [Thread-632 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:58,804 WARN [Thread-612 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:13:58,810 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x22414377bfd79a8c with lease ID 0xeb533680bbc0558d: Processing first storage report for DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd from datanode DatanodeRegistration(127.0.0.1:35103, datanodeUuid=9e57a236-2e7d-457a-9a0b-d39386031171, infoPort=43925, infoSecurePort=0, ipcPort=45709, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:58,811 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x22414377bfd79a8c with lease ID 0xeb533680bbc0558d: from storage DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd node DatanodeRegistration(127.0.0.1:35103, datanodeUuid=9e57a236-2e7d-457a-9a0b-d39386031171, infoPort=43925, infoSecurePort=0, ipcPort=45709, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:58,811 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x22414377bfd79a8c with lease ID 0xeb533680bbc0558d: Processing first storage report for DS-70f3f200-c4df-4d86-ae32-82f219ff4199 from datanode DatanodeRegistration(127.0.0.1:35103, datanodeUuid=9e57a236-2e7d-457a-9a0b-d39386031171, infoPort=43925, infoSecurePort=0, ipcPort=45709, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:58,811 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x22414377bfd79a8c with lease ID 0xeb533680bbc0558d: from storage DS-70f3f200-c4df-4d86-ae32-82f219ff4199 node DatanodeRegistration(127.0.0.1:35103, datanodeUuid=9e57a236-2e7d-457a-9a0b-d39386031171, infoPort=43925, infoSecurePort=0, ipcPort=45709, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:58,880 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@55b30d58{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir/jetty-localhost-34531-hadoop-hdfs-3_4_1-tests_jar-_-any-3822542457635880649/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:58,880 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3d217d2a{HTTP/1.1, (http/1.1)}{localhost:34531} 2024-12-05T15:13:58,880 INFO [Time-limited test {}] server.Server(415): Started @129696ms 2024-12-05T15:13:58,882 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:13:58,916 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:13:58,922 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:13:58,930 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:13:58,930 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:13:58,930 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:13:58,931 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6a9b440a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:13:58,931 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@186aeb49{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:13:58,986 WARN [Thread-668 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data8/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:58,986 WARN [Thread-667 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data7/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:59,004 WARN [Thread-647 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:13:59,008 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb28221cd2c13ff29 with lease ID 0xeb533680bbc0558e: Processing first storage report for DS-e099b2ba-f716-4289-a26f-15463d9b7d5b from datanode DatanodeRegistration(127.0.0.1:42629, datanodeUuid=9b637737-2713-4915-84e1-e2c83acc45ce, infoPort=43703, infoSecurePort=0, ipcPort=43125, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:59,008 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb28221cd2c13ff29 with lease ID 0xeb533680bbc0558e: from storage DS-e099b2ba-f716-4289-a26f-15463d9b7d5b node DatanodeRegistration(127.0.0.1:42629, datanodeUuid=9b637737-2713-4915-84e1-e2c83acc45ce, infoPort=43703, infoSecurePort=0, ipcPort=43125, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:59,008 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xb28221cd2c13ff29 with lease ID 0xeb533680bbc0558e: Processing first storage report for DS-37ff94bb-6b54-4b1f-9a54-9c8994d21ac1 from datanode DatanodeRegistration(127.0.0.1:42629, datanodeUuid=9b637737-2713-4915-84e1-e2c83acc45ce, infoPort=43703, infoSecurePort=0, ipcPort=43125, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:59,008 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xb28221cd2c13ff29 with lease ID 0xeb533680bbc0558e: from storage DS-37ff94bb-6b54-4b1f-9a54-9c8994d21ac1 node DatanodeRegistration(127.0.0.1:42629, datanodeUuid=9b637737-2713-4915-84e1-e2c83acc45ce, infoPort=43703, infoSecurePort=0, ipcPort=43125, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:59,049 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@491ae7dc{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir/jetty-localhost-43695-hadoop-hdfs-3_4_1-tests_jar-_-any-7627263902710637907/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:59,049 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@e0d0b01{HTTP/1.1, (http/1.1)}{localhost:43695} 2024-12-05T15:13:59,049 INFO [Time-limited test {}] server.Server(415): Started @129865ms 2024-12-05T15:13:59,051 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:13:59,150 WARN [Thread-694 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data10/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:59,150 WARN [Thread-693 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data9/current/BP-308116525-172.17.0.2-1733411625957/current, will proceed with Du for space computation calculation, 2024-12-05T15:13:59,168 WARN [Thread-682 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:13:59,170 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xaa494917e1a3526b with lease ID 0xeb533680bbc0558f: Processing first storage report for DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc from datanode DatanodeRegistration(127.0.0.1:40345, datanodeUuid=fc78e194-26d6-45c1-ae67-cbd211cc42db, infoPort=42747, infoSecurePort=0, ipcPort=44879, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:59,171 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xaa494917e1a3526b with lease ID 0xeb533680bbc0558f: from storage DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc node DatanodeRegistration(127.0.0.1:40345, datanodeUuid=fc78e194-26d6-45c1-ae67-cbd211cc42db, infoPort=42747, infoSecurePort=0, ipcPort=44879, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:59,171 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xaa494917e1a3526b with lease ID 0xeb533680bbc0558f: Processing first storage report for DS-ed29022b-439c-4340-9fdd-128a0d2e288d from datanode DatanodeRegistration(127.0.0.1:40345, datanodeUuid=fc78e194-26d6-45c1-ae67-cbd211cc42db, infoPort=42747, infoSecurePort=0, ipcPort=44879, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957) 2024-12-05T15:13:59,171 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xaa494917e1a3526b with lease ID 0xeb533680bbc0558f: from storage DS-ed29022b-439c-4340-9fdd-128a0d2e288d node DatanodeRegistration(127.0.0.1:40345, datanodeUuid=fc78e194-26d6-45c1-ae67-cbd211cc42db, infoPort=42747, infoSecurePort=0, ipcPort=44879, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:13:59,174 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1010 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1010 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,174 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,175 WARN [DataStreamer for file /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 block BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009 in pipeline [DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:13:59,175 WARN [DataStreamer for file /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta block BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1010 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1010 in pipeline [DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:13:59,174 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1006 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1006 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,175 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015 java.io.IOException: Bad response ERROR for BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015 from datanode DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,176 WARN [DataStreamer for file /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 block BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1006 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1006 in pipeline [DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:13:59,179 WARN [DataStreamer for file /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,39169,1733411628416/4ecb0f57b706%2C39169%2C1733411628416.1733411628627 block BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK], DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:13:59,180 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@2d70ba54{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:59,179 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1231261664_22 at /127.0.0.1:48314 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48314 dst: /127.0.0.1:43797 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,179 WARN [PacketResponder: BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:41001] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,179 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49036 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:41001:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49036 dst: /127.0.0.1:41001 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,179 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1231261664_22 at /127.0.0.1:49002 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:41001:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49002 dst: /127.0.0.1:41001 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,181 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1a6c4f80{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:13:59,181 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:13:59,179 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49052 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:41001:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49052 dst: /127.0.0.1:41001 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,181 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@60b15398{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:13:59,181 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1885803086_22 at /127.0.0.1:48416 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015] {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48416 dst: /127.0.0.1:43797 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,182 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@29f7fced{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,STOPPED} 2024-12-05T15:13:59,180 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:48348 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48348 dst: /127.0.0.1:43797 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,182 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1885803086_22 at /127.0.0.1:49102 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015] {}] datanode.DataXceiver(331): 127.0.0.1:41001:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49102 dst: /127.0.0.1:41001 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,183 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:48354 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:48354 dst: /127.0.0.1:43797 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,185 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:13:59,185 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:13:59,185 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:13:59,185 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-308116525-172.17.0.2-1733411625957 (Datanode Uuid aee4fce9-277a-43d3-9380-0e0a2c5bf5c9) service to localhost/127.0.0.1:32987 2024-12-05T15:13:59,186 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data3/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:59,186 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data4/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:59,186 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:13:59,188 ERROR [org.apache.hadoop.hdfs.server.datanode.DataXceiver@29840957 {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing unknown operation src: /127.0.0.1:59770 dst: /127.0.0.1:43797 java.io.IOException: Server closed. at org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.addPeer(DataXceiverServer.java:334) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:232) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,188 WARN [DataStreamer for file /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 block BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741833_1009 java.net.SocketException: Connection reset at sun.nio.ch.SocketChannelImpl.throwConnectionReset(SocketChannelImpl.java:394) ~[?:?] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:426) ~[?:?] at org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:118) ~[hadoop-common-3.4.1.jar:?] at java.io.FilterInputStream.read(FilterInputStream.java:82) ~[?:?] at java.io.FilterInputStream.read(FilterInputStream.java:82) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:527) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1931) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1666) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,192 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1885803086_22 at /127.0.0.1:59768 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1015] {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59768 dst: /127.0.0.1:43797 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,192 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1231261664_22 at /127.0.0.1:59774 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59774 dst: /127.0.0.1:43797 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,193 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:59772 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:43797:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:59772 dst: /127.0.0.1:43797 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:13:59,194 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1018 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741834_1018 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,194 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1016 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741830_1016 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,195 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1019 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741839_1019 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,196 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5ecf65b2{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:13:59,197 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5483975{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:13:59,197 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:13:59,197 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@220b71ae{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:13:59,197 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@59774852{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,STOPPED} 2024-12-05T15:13:59,198 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:13:59,198 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:13:59,198 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:13:59,198 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-308116525-172.17.0.2-1733411625957 (Datanode Uuid 827e5892-6c93-4960-ae7d-59bc773f1ed4) service to localhost/127.0.0.1:32987 2024-12-05T15:13:59,199 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data1/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:59,199 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data2/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:13:59,200 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:13:59,204 WARN [RS:0;4ecb0f57b706:44787.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=4, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,204 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C44787%2C1733411626859:(num 1733411627293) roll requested 2024-12-05T15:13:59,205 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.1733411639205 2024-12-05T15:13:59,205 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=44787 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=4, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=44787 {}] ipc.CallRunner(138): callId: 9 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:57764 deadline: 1733411649204, exception=org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=4, requesting roll of WAL 2024-12-05T15:13:59,211 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=4, requesting roll of WAL 2024-12-05T15:13:59,211 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 with entries=4, filesize=959 B; new WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411639205 2024-12-05T15:13:59,211 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43703:43703),(127.0.0.1/127.0.0.1:42747:42747)] 2024-12-05T15:13:59,211 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 is not closed yet, will try archiving it next time 2024-12-05T15:13:59,211 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,211 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:13:59,212 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(47): Initialize RecoverLeaseFSUtils 2024-12-05T15:13:59,212 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(59): set recoverLeaseMethod to org.apache.hadoop.fs.LeaseRecoverable.recoverLease() 2024-12-05T15:13:59,212 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 2024-12-05T15:13:59,215 WARN [IPC Server handler 0 on default port 32987 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 has not been closed. Lease recovery is in progress. RecoveryId = 1021 for block blk_1073741833_1009 2024-12-05T15:13:59,217 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 after 5ms 2024-12-05T15:14:00,496 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:14:00,497 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:00,525 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:00,528 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:00,528 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:03,218 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=1 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 after 4006ms 2024-12-05T15:14:11,240 INFO [Time-limited test {}] wal.TestLogRolling(243): log.getCurrentFileName(): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411639205 2024-12-05T15:14:11,240 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741840_1020 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741840_1020 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:11,241 WARN [DataStreamer for file /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411639205 block BP-308116525-172.17.0.2-1733411625957:blk_1073741840_1020 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741840_1020 in pipeline [DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK], DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]) is bad. 2024-12-05T15:14:11,241 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:34230 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741840_1020] {}] datanode.DataXceiver(331): 127.0.0.1:42629:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:34230 dst: /127.0.0.1:42629 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:11,242 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:39200 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741840_1020] {}] datanode.DataXceiver(331): 127.0.0.1:40345:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39200 dst: /127.0.0.1:40345 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:11,243 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@55b30d58{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:14:11,244 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3d217d2a{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:14:11,244 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:14:11,244 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@55167cef{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:14:11,244 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@493e02d8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,STOPPED} 2024-12-05T15:14:11,247 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:14:11,247 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:14:11,247 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-308116525-172.17.0.2-1733411625957 (Datanode Uuid 9b637737-2713-4915-84e1-e2c83acc45ce) service to localhost/127.0.0.1:32987 2024-12-05T15:14:11,247 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:14:11,247 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data7/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:11,247 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data8/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:11,248 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:14:11,250 WARN [sync.1 {}] wal.FSHLog(750): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]] 2024-12-05T15:14:11,250 WARN [sync.1 {}] wal.FSHLog(721): Requesting log roll because of low replication, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]] 2024-12-05T15:14:11,250 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C44787%2C1733411626859:(num 1733411639205) roll requested 2024-12-05T15:14:11,250 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.1733411651250 2024-12-05T15:14:11,254 WARN [Thread-724 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741841_1023 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:43797 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:11,254 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49428 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741841_1023] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6]'}, localName='127.0.0.1:35103', datanodeUuid='9e57a236-2e7d-457a-9a0b-d39386031171', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741841_1023 to mirror 127.0.0.1:43797 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:11,254 WARN [Thread-724 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741841_1023 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]) is bad. 2024-12-05T15:14:11,254 WARN [Thread-724 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741841_1023 2024-12-05T15:14:11,254 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49428 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741841_1023] {}] datanode.BlockReceiver(316): Block 1073741841 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-05T15:14:11,255 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49428 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741841_1023] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49428 dst: /127.0.0.1:35103 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:11,257 WARN [Thread-724 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK] 2024-12-05T15:14:11,260 WARN [Thread-724 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741842_1024 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:11,260 WARN [Thread-724 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741842_1024 in pipeline [DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK], DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:14:11,260 WARN [Thread-724 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741842_1024 2024-12-05T15:14:11,261 WARN [Thread-724 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK] 2024-12-05T15:14:11,263 WARN [Thread-724 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741843_1025 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:42629 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:11,263 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49436 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741843_1025] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6]'}, localName='127.0.0.1:35103', datanodeUuid='9e57a236-2e7d-457a-9a0b-d39386031171', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741843_1025 to mirror 127.0.0.1:42629 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:11,263 WARN [Thread-724 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741843_1025 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]) is bad. 2024-12-05T15:14:11,263 WARN [Thread-724 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741843_1025 2024-12-05T15:14:11,263 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49436 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741843_1025] {}] datanode.BlockReceiver(316): Block 1073741843 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-05T15:14:11,264 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49436 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741843_1025] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49436 dst: /127.0.0.1:35103 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:11,264 WARN [Thread-724 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK] 2024-12-05T15:14:11,269 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411639205 with entries=2, filesize=2.37 KB; new WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411651250 2024-12-05T15:14:11,272 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43925:43925),(127.0.0.1/127.0.0.1:42747:42747)] 2024-12-05T15:14:11,272 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 is not closed yet, will try archiving it next time 2024-12-05T15:14:11,272 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411639205 is not closed yet, will try archiving it next time 2024-12-05T15:14:11,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:40345 is added to blk_1073741840_1022 (size=2431) 2024-12-05T15:14:11,674 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 is not closed yet, will try archiving it next time 2024-12-05T15:14:14,181 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@647d7b64[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:40345, datanodeUuid=fc78e194-26d6-45c1-ae67-cbd211cc42db, infoPort=42747, infoSecurePort=0, ipcPort=44879, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741840_1022 to 127.0.0.1:41001 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,254 WARN [ResponseProcessor for block BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026 java.io.IOException: Bad response ERROR for BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026 from datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,254 WARN [DataStreamer for file /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411651250 block BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:15,254 WARN [PacketResponder: BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:40345] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,255 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:49448 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:49448 dst: /127.0.0.1:35103 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,255 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:36480 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1026] {}] datanode.DataXceiver(331): 127.0.0.1:40345:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:36480 dst: /127.0.0.1:40345 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,257 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@491ae7dc{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:14:15,257 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@e0d0b01{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:14:15,257 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:14:15,257 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@186aeb49{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:14:15,258 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6a9b440a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,STOPPED} 2024-12-05T15:14:15,259 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:14:15,259 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:14:15,259 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:14:15,259 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-308116525-172.17.0.2-1733411625957 (Datanode Uuid fc78e194-26d6-45c1-ae67-cbd211cc42db) service to localhost/127.0.0.1:32987 2024-12-05T15:14:15,260 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data9/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:15,260 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data10/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:15,261 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:14:15,263 WARN [sync.4 {}] wal.FSHLog(750): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK]] 2024-12-05T15:14:15,263 WARN [sync.4 {}] wal.FSHLog(721): Requesting log roll because of low replication, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK]] 2024-12-05T15:14:15,263 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C44787%2C1733411626859:(num 1733411651250) roll requested 2024-12-05T15:14:15,264 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.1733411655263 2024-12-05T15:14:15,266 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741845_1028 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,267 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741845_1028 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK], DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]) is bad. 2024-12-05T15:14:15,267 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741845_1028 2024-12-05T15:14:15,267 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK] 2024-12-05T15:14:15,268 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=44787 {}] regionserver.HRegion(8581): Flush requested on 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:14:15,268 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 817a20dfbf0d1531445f95325bcaf27c 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:14:15,270 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741846_1029 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:42629 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,270 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46270 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741846_1029] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6]'}, localName='127.0.0.1:35103', datanodeUuid='9e57a236-2e7d-457a-9a0b-d39386031171', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741846_1029 to mirror 127.0.0.1:42629 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,271 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741846_1029 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]) is bad. 2024-12-05T15:14:15,271 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741846_1029 2024-12-05T15:14:15,271 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46270 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741846_1029] {}] datanode.BlockReceiver(316): Block 1073741846 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-05T15:14:15,271 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46270 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741846_1029] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46270 dst: /127.0.0.1:35103 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,271 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK] 2024-12-05T15:14:15,273 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741847_1030 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,273 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741847_1030 in pipeline [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK], DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:15,273 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741847_1030 2024-12-05T15:14:15,274 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:15,277 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46280 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741848_1031] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6]'}, localName='127.0.0.1:35103', datanodeUuid='9e57a236-2e7d-457a-9a0b-d39386031171', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741848_1031 to mirror 127.0.0.1:41001 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,277 WARN [Thread-738 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741848_1031 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:41001 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,277 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46280 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741848_1031] {}] datanode.BlockReceiver(316): Block 1073741848 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-05T15:14:15,277 WARN [Thread-738 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741848_1031 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:14:15,277 WARN [Thread-738 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741848_1031 2024-12-05T15:14:15,277 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46280 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741848_1031] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46280 dst: /127.0.0.1:35103 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,278 WARN [Thread-738 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK] 2024-12-05T15:14:15,279 WARN [IPC Server handler 3 on default port 32987 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-05T15:14:15,279 WARN [IPC Server handler 3 on default port 32987 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-05T15:14:15,279 WARN [IPC Server handler 3 on default port 32987 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-05T15:14:15,283 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411651250 with entries=9, filesize=9.53 KB; new WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655263 2024-12-05T15:14:15,283 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43925:43925)] 2024-12-05T15:14:15,283 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 is not closed yet, will try archiving it next time 2024-12-05T15:14:15,283 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411651250 is not closed yet, will try archiving it next time 2024-12-05T15:14:15,283 WARN [sync.3 {}] wal.FSHLog(750): HDFS pipeline error detected. Found 1 replicas but expecting no less than 2 replicas. Requesting close of WAL. current pipeline: [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK]] 2024-12-05T15:14:15,283 WARN [sync.3 {}] wal.FSHLog(721): Requesting log roll because of low replication, current pipeline: [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK]] 2024-12-05T15:14:15,283 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C44787%2C1733411626859:(num 1733411655263) roll requested 2024-12-05T15:14:15,284 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.1733411655284 2024-12-05T15:14:15,285 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741844_1027 (size=9768) 2024-12-05T15:14:15,286 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 is not closed yet, will try archiving it next time 2024-12-05T15:14:15,287 WARN [Thread-745 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741850_1033 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,287 WARN [Thread-745 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741850_1033 in pipeline [DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK], DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:14:15,287 WARN [Thread-745 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741850_1033 2024-12-05T15:14:15,288 WARN [Thread-745 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK] 2024-12-05T15:14:15,289 WARN [Thread-745 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741851_1034 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,289 WARN [Thread-745 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741851_1034 in pipeline [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:15,289 WARN [Thread-745 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741851_1034 2024-12-05T15:14:15,290 WARN [Thread-745 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:15,291 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/.tmp/info/94705e506d88441ca474b02cfc8a47be is 1080, key is row0002/info:/1733411651249/Put/seqid=0 2024-12-05T15:14:15,292 WARN [Thread-745 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741852_1035 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:43797 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,292 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46302 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741852_1035] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6]'}, localName='127.0.0.1:35103', datanodeUuid='9e57a236-2e7d-457a-9a0b-d39386031171', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741852_1035 to mirror 127.0.0.1:43797 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,292 WARN [Thread-745 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741852_1035 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]) is bad. 2024-12-05T15:14:15,292 WARN [Thread-745 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741852_1035 2024-12-05T15:14:15,292 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46302 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741852_1035] {}] datanode.BlockReceiver(316): Block 1073741852 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-05T15:14:15,292 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46302 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741852_1035] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46302 dst: /127.0.0.1:35103 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,292 WARN [Thread-740 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741853_1036 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,293 WARN [Thread-740 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741853_1036 in pipeline [DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]) is bad. 2024-12-05T15:14:15,293 WARN [Thread-745 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK] 2024-12-05T15:14:15,293 WARN [Thread-740 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741853_1036 2024-12-05T15:14:15,293 WARN [Thread-740 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:41001,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK] 2024-12-05T15:14:15,294 WARN [Thread-745 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741854_1037 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,294 WARN [Thread-745 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741854_1037 in pipeline [DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK], DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]) is bad. 2024-12-05T15:14:15,294 WARN [Thread-745 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741854_1037 2024-12-05T15:14:15,294 WARN [Thread-740 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741855_1038 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,294 WARN [Thread-740 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741855_1038 in pipeline [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK], DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:15,294 WARN [Thread-740 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741855_1038 2024-12-05T15:14:15,295 WARN [Thread-745 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK] 2024-12-05T15:14:15,295 WARN [Thread-740 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:15,295 WARN [IPC Server handler 1 on default port 32987 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-05T15:14:15,295 WARN [IPC Server handler 1 on default port 32987 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-05T15:14:15,296 WARN [IPC Server handler 1 on default port 32987 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-05T15:14:15,296 WARN [Thread-740 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741856_1039 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,296 WARN [Thread-740 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741856_1039 in pipeline [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK], DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]) is bad. 2024-12-05T15:14:15,296 WARN [Thread-740 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741856_1039 2024-12-05T15:14:15,297 WARN [Thread-740 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK] 2024-12-05T15:14:15,301 WARN [Thread-740 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741858_1041 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:42629 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:15,301 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46316 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741858_1041] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6]'}, localName='127.0.0.1:35103', datanodeUuid='9e57a236-2e7d-457a-9a0b-d39386031171', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741858_1041 to mirror 127.0.0.1:42629 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,302 WARN [Thread-740 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741858_1041 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK]) is bad. 2024-12-05T15:14:15,302 WARN [Thread-740 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741858_1041 2024-12-05T15:14:15,302 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46316 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741858_1041] {}] datanode.BlockReceiver(316): Block 1073741858 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-05T15:14:15,302 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:46316 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741858_1041] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:46316 dst: /127.0.0.1:35103 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:15,302 WARN [Thread-740 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:42629,DS-e099b2ba-f716-4289-a26f-15463d9b7d5b,DISK] 2024-12-05T15:14:15,303 WARN [IPC Server handler 0 on default port 32987 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) For more information, please enable DEBUG log level on org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy and org.apache.hadoop.net.NetworkTopology 2024-12-05T15:14:15,303 WARN [IPC Server handler 0 on default port 32987 {}] protocol.BlockStoragePolicy(161): Failed to place enough replicas: expected size is 1 but only 0 storage types can be selected (replication=2, selected=[], unavailable=[DISK], removed=[DISK], policy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}) 2024-12-05T15:14:15,303 WARN [IPC Server handler 0 on default port 32987 {}] blockmanagement.BlockPlacementPolicyDefault(501): Failed to place enough replicas, still in need of 1 to reach 2 (unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]}, newBlock=true) All required storage types are unavailable: unavailableStorages=[DISK], storagePolicy=BlockStoragePolicy{HOT:7, storageTypes=[DISK], creationFallbacks=[], replicationFallbacks=[ARCHIVE]} 2024-12-05T15:14:15,304 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655263 with entries=4, filesize=4.65 KB; new WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655284 2024-12-05T15:14:15,304 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43925:43925)] 2024-12-05T15:14:15,304 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 is not closed yet, will try archiving it next time 2024-12-05T15:14:15,305 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655263 is not closed yet, will try archiving it next time 2024-12-05T15:14:15,306 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741849_1032 (size=4768) 2024-12-05T15:14:15,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741859_1042 (size=10347) 2024-12-05T15:14:15,313 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=12 (bloomFilter=true), to=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/.tmp/info/94705e506d88441ca474b02cfc8a47be 2024-12-05T15:14:15,322 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/.tmp/info/94705e506d88441ca474b02cfc8a47be as hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/info/94705e506d88441ca474b02cfc8a47be 2024-12-05T15:14:15,328 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/info/94705e506d88441ca474b02cfc8a47be, entries=5, sequenceid=12, filesize=10.1 K 2024-12-05T15:14:15,329 WARN [sync.3 {}] wal.FSHLog(760): Too many consecutive RollWriter requests, it's a sign of the total number of live datanodes is lower than the tolerable replicas. 2024-12-05T15:14:15,329 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=7.35 KB/7531 for 817a20dfbf0d1531445f95325bcaf27c in 61ms, sequenceid=12, compaction requested=false 2024-12-05T15:14:15,329 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 817a20dfbf0d1531445f95325bcaf27c: 2024-12-05T15:14:15,500 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:14:15,504 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:14:15,505 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:14:15,505 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:14:15,505 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:14:15,506 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@20e478b8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:14:15,506 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@601bb964{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:14:15,622 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@7d38b2e7{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/java.io.tmpdir/jetty-localhost-45029-hadoop-hdfs-3_4_1-tests_jar-_-any-10070817255352361103/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:14:15,622 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1783a9ee{HTTP/1.1, (http/1.1)}{localhost:45029} 2024-12-05T15:14:15,622 INFO [Time-limited test {}] server.Server(415): Started @146438ms 2024-12-05T15:14:15,624 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:14:15,707 DEBUG [Close-WAL-Writer-1 {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 is not closed yet, will try archiving it next time 2024-12-05T15:14:15,708 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411639205 to hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs/4ecb0f57b706%2C44787%2C1733411626859.1733411639205 2024-12-05T15:14:15,755 WARN [Thread-769 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:14:15,763 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa33bd52521e7d2da with lease ID 0xeb533680bbc05590: from storage DS-6781d264-5b19-4912-a89b-9a138120ec6a node DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 8, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-05T15:14:15,763 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xa33bd52521e7d2da with lease ID 0xeb533680bbc05590: from storage DS-39ac30c5-c7cf-427d-8a5c-40fc28623f09 node DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:14:16,781 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:14:16,813 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@786e5e3d[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:35103, datanodeUuid=9e57a236-2e7d-457a-9a0b-d39386031171, infoPort=43925, infoSecurePort=0, ipcPort=45709, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741849_1032 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:16,813 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@155d7f95[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:35103, datanodeUuid=9e57a236-2e7d-457a-9a0b-d39386031171, infoPort=43925, infoSecurePort=0, ipcPort=45709, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741844_1027 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:17,067 WARN [master/4ecb0f57b706:0:becomeActiveMaster.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=96, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:17,067 DEBUG [master:store-WAL-Roller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C38149%2C1733411626804:(num 1733411626978) roll requested 2024-12-05T15:14:17,068 INFO [master:store-WAL-Roller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C38149%2C1733411626804.1733411657068 2024-12-05T15:14:17,068 ERROR [ProcExecTimeout {}] region.RegionProcedureStore(422): Failed to delete pids=[4, 7, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:17,068 ERROR [ProcExecTimeout {}] procedure2.TimeoutExecutorThread(124): Ignoring pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner exception: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL java.io.UncheckedIOException: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL at org.apache.hadoop.hbase.procedure2.store.region.RegionProcedureStore.delete(RegionProcedureStore.java:423) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner.periodicExecute(CompletedProcedureCleaner.java:135) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.executeInMemoryChore(TimeoutExecutorThread.java:122) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.execDelayedProcedure(TimeoutExecutorThread.java:101) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.run(TimeoutExecutorThread.java:68) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] Caused by: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:17,071 WARN [Thread-791 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741860_1043 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:17,071 WARN [Thread-791 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741860_1043 in pipeline [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK], DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:17,072 WARN [Thread-791 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741860_1043 2024-12-05T15:14:17,072 WARN [Thread-791 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:17,078 WARN [master:store-WAL-Roller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=96, requesting roll of WAL 2024-12-05T15:14:17,078 INFO [master:store-WAL-Roller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 with entries=93, filesize=46.05 KB; new WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411657068 2024-12-05T15:14:17,079 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:43925:43925),(127.0.0.1/127.0.0.1:45209:45209)] 2024-12-05T15:14:17,079 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 is not closed yet, will try archiving it next time 2024-12-05T15:14:17,079 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:17,079 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:17,079 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 2024-12-05T15:14:17,080 WARN [IPC Server handler 3 on default port 32987 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 has not been closed. Lease recovery is in progress. RecoveryId = 1045 for block blk_1073741830_1016 2024-12-05T15:14:17,080 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 after 1ms 2024-12-05T15:14:17,813 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@155d7f95[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:35103, datanodeUuid=9e57a236-2e7d-457a-9a0b-d39386031171, infoPort=43925, infoSecurePort=0, ipcPort=45709, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741859_1042 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:20,453 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:14:20,455 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50258, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:14:21,082 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=1 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 after 4002ms 2024-12-05T15:14:22,112 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:14:22,113 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:50272, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:14:25,778 WARN [org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1@6c4fc442 {}] datanode.BlockRecoveryWorker$RecoveryTaskContiguous(164): Failed to recover block (block=BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1009, datanode=DatanodeInfoWithStorage[127.0.0.1:43797,null,null]) java.net.ConnectException: Call From 4ecb0f57b706/172.17.0.2 to localhost:44937 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:961) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:876) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1588) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy55.initReplicaRecovery(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.lambda$initReplicaRecovery$0(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.InterDatanodeProtocolTranslatorPB.initReplicaRecovery(InterDatanodeProtocolTranslatorPB.java:82) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.callInitReplicaRecovery(BlockRecoveryWorker.java:561) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker.access$400(BlockRecoveryWorker.java:57) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$RecoveryTaskContiguous.recover(BlockRecoveryWorker.java:135) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockRecoveryWorker$1.run(BlockRecoveryWorker.java:602) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:668) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:789) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client$Connection.access$3800(Client.java:364) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.getConnection(Client.java:1649) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1473) ~[hadoop-common-3.4.1.jar:?] ... 12 more 2024-12-05T15:14:25,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741833_1021 (size=959) 2024-12-05T15:14:28,410 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-05T15:14:28,410 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-05T15:14:28,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@66d38368[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741833_1021 to 127.0.0.1:42629 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:28,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@5d6988cd[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741838_1014 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:29,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@5d6988cd[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741836_1012 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:29,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@66d38368[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741832_1008 to 127.0.0.1:42629 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:31,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@66d38368[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741831_1007 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:31,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@5d6988cd[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741829_1005 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:32,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@66d38368[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741825_1001 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:32,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@5d6988cd[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741827_1003 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:33,881 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 817a20dfbf0d1531445f95325bcaf27c, had cached 0 bytes from a total of 10347 2024-12-05T15:14:34,758 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@5d6988cd[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741828_1004 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:34,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:14:35,204 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.1733411675204 2024-12-05T15:14:35,211 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655284 with entries=1, filesize=442 B; new WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411675204 2024-12-05T15:14:35,211 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45209:45209),(127.0.0.1/127.0.0.1:43925:43925)] 2024-12-05T15:14:35,211 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655284 is not closed yet, will try archiving it next time 2024-12-05T15:14:35,213 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741857_1040 (size=450) 2024-12-05T15:14:35,213 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=44787 {}] regionserver.HRegion(8581): Flush requested on 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:14:35,214 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 817a20dfbf0d1531445f95325bcaf27c 1/1 column families, dataSize=8.41 KB heapSize=9.25 KB 2024-12-05T15:14:35,215 INFO [sync.1 {}] wal.FSHLog(777): LowReplication-Roller was enabled. 2024-12-05T15:14:35,219 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/.tmp/info/3f8e3d1914ba4826afaca2ef4c5299da is 1080, key is row0007/info:/1733411655269/Put/seqid=0 2024-12-05T15:14:35,221 WARN [Thread-807 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741863_1047 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,221 WARN [Thread-807 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741863_1047 in pipeline [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK], DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:35,221 WARN [Thread-807 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741863_1047 2024-12-05T15:14:35,221 WARN [Thread-807 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:35,227 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-05T15:14:35,227 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-05T15:14:35,227 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6fb20afc to 127.0.0.1:55344 2024-12-05T15:14:35,227 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:14:35,227 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T15:14:35,227 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1904488592, stopped=false 2024-12-05T15:14:35,227 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=4ecb0f57b706,38149,1733411626804 2024-12-05T15:14:35,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741864_1048 (size=13583) 2024-12-05T15:14:35,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741864_1048 (size=13583) 2024-12-05T15:14:35,230 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=8.41 KB at sequenceid=23 (bloomFilter=true), to=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/.tmp/info/3f8e3d1914ba4826afaca2ef4c5299da 2024-12-05T15:14:35,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:14:35,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:14:35,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:35,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:35,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:14:35,232 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-05T15:14:35,232 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:35,232 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:14:35,232 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,44787,1733411626859' ***** 2024-12-05T15:14:35,232 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-05T15:14:35,232 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,39169,1733411628416' ***** 2024-12-05T15:14:35,232 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-05T15:14:35,232 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T15:14:35,232 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T15:14:35,232 INFO [RS:1;4ecb0f57b706:39169 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T15:14:35,232 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-05T15:14:35,232 INFO [RS:1;4ecb0f57b706:39169 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T15:14:35,233 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,39169,1733411628416 2024-12-05T15:14:35,233 DEBUG [RS:1;4ecb0f57b706:39169 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:14:35,233 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,39169,1733411628416; all regions closed. 2024-12-05T15:14:35,233 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:14:35,233 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:14:35,233 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:14:35,233 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,39169,1733411628416 2024-12-05T15:14:35,233 WARN [WAL-Shutdown-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,234 ERROR [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1664): Shutdown / close of WAL failed: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... 2024-12-05T15:14:35,234 DEBUG [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1665): Shutdown / close exception details: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,234 DEBUG [RS:1;4ecb0f57b706:39169 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:14:35,234 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:14:35,234 INFO [RS:1;4ecb0f57b706:39169 {}] hbase.ChoreService(370): Chore service for: regionserver/4ecb0f57b706:0 had [ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS] on shutdown 2024-12-05T15:14:35,234 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:14:35,234 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T15:14:35,235 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T15:14:35,235 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T15:14:35,235 INFO [RS:1;4ecb0f57b706:39169 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:39169 2024-12-05T15:14:35,236 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/4ecb0f57b706,39169,1733411628416 2024-12-05T15:14:35,236 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:14:35,237 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [4ecb0f57b706,39169,1733411628416] 2024-12-05T15:14:35,237 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 4ecb0f57b706,39169,1733411628416; numProcessing=1 2024-12-05T15:14:35,239 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/4ecb0f57b706,39169,1733411628416 already deleted, retry=false 2024-12-05T15:14:35,239 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 4ecb0f57b706,39169,1733411628416 expired; onlineServers=1 2024-12-05T15:14:35,239 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/.tmp/info/3f8e3d1914ba4826afaca2ef4c5299da as hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/info/3f8e3d1914ba4826afaca2ef4c5299da 2024-12-05T15:14:35,245 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/info/3f8e3d1914ba4826afaca2ef4c5299da, entries=8, sequenceid=23, filesize=13.3 K 2024-12-05T15:14:35,246 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~8.41 KB/8607, heapSize ~9.23 KB/9456, currentSize=9.46 KB/9684 for 817a20dfbf0d1531445f95325bcaf27c in 32ms, sequenceid=23, compaction requested=false 2024-12-05T15:14:35,246 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 817a20dfbf0d1531445f95325bcaf27c: 2024-12-05T15:14:35,246 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=23.4 K, sizeToCheck=16.0 K 2024-12-05T15:14:35,247 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:14:35,247 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/default/TestLogRolling-testLogRollOnDatanodeDeath/817a20dfbf0d1531445f95325bcaf27c/info/3f8e3d1914ba4826afaca2ef4c5299da because midkey is the same as first or last row 2024-12-05T15:14:35,247 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(3579): Received CLOSE for 474a835416f2f984a3141295bfac95bd 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(3579): Received CLOSE for 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,44787,1733411626859 2024-12-05T15:14:35,247 DEBUG [RS:0;4ecb0f57b706:44787 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T15:14:35,247 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 474a835416f2f984a3141295bfac95bd, disabling compactions & flushes 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T15:14:35,247 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:14:35,247 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:14:35,247 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:14:35,248 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. after waiting 0 ms 2024-12-05T15:14:35,248 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:14:35,248 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 474a835416f2f984a3141295bfac95bd 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-05T15:14:35,248 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-05T15:14:35,248 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1603): Online Regions={474a835416f2f984a3141295bfac95bd=hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd., 1588230740=hbase:meta,,1.1588230740, 817a20dfbf0d1531445f95325bcaf27c=TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c.} 2024-12-05T15:14:35,248 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:14:35,248 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 474a835416f2f984a3141295bfac95bd, 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:14:35,248 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:14:35,248 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:14:35,248 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:14:35,248 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:14:35,248 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.87 KB heapSize=5.40 KB 2024-12-05T15:14:35,249 WARN [RS_OPEN_META-regionserver/4ecb0f57b706:0-0.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=15, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,249 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C44787%2C1733411626859.meta:.meta(num 1733411627726) roll requested 2024-12-05T15:14:35,249 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:14:35,249 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C44787%2C1733411626859.meta.1733411675249.meta 2024-12-05T15:14:35,249 ERROR [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2808): ***** ABORTING region server 4ecb0f57b706,44787,1733411626859: Unrecoverable exception while closing hbase:meta,,1.1588230740 ***** org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,250 ERROR [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2815): RegionServer abort: loaded coprocessors are: [org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint] 2024-12-05T15:14:35,253 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for java.lang:type=Memory 2024-12-05T15:14:35,253 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:45986 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741865_1049] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data3, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data4]'}, localName='127.0.0.1:32933', datanodeUuid='aee4fce9-277a-43d3-9380-0e0a2c5bf5c9', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741865_1049 to mirror 127.0.0.1:40345 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:35,253 WARN [Thread-815 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741865_1049 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:40345 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,253 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:45986 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741865_1049] {}] datanode.BlockReceiver(316): Block 1073741865 has not released the reserved bytes. Releasing 268435456 bytes as part of close. 2024-12-05T15:14:35,253 WARN [Thread-815 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741865_1049 in pipeline [DatanodeInfoWithStorage[127.0.0.1:32933,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK], DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:35,253 WARN [Thread-815 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741865_1049 2024-12-05T15:14:35,253 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-111281336_22 at /127.0.0.1:45986 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741865_1049] {}] datanode.DataXceiver(331): 127.0.0.1:32933:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45986 dst: /127.0.0.1:32933 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:35,254 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=IPC 2024-12-05T15:14:35,254 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=Replication 2024-12-05T15:14:35,255 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=Server 2024-12-05T15:14:35,255 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2819): Dump of metrics as JSON on abort: { "beans": [ { "name": "java.lang:type=Memory", "modelerType": "sun.management.MemoryImpl", "ObjectPendingFinalizationCount": 0, "HeapMemoryUsage": { "committed": 1048576000, "init": 1048576000, "max": 2306867200, "used": 281724400 }, "NonHeapMemoryUsage": { "committed": 162725888, "init": 7667712, "max": -1, "used": 160888416 }, "Verbose": false, "ObjectName": "java.lang:type=Memory" } ], "beans": [], "beans": [], "beans": [] } 2024-12-05T15:14:35,258 WARN [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=38149 {}] master.MasterRpcServices(626): 4ecb0f57b706,44787,1733411626859 reported a fatal error: ***** ABORTING region server 4ecb0f57b706,44787,1733411626859: Unrecoverable exception while closing hbase:meta,,1.1588230740 ***** Cause: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) at java.base/java.lang.Thread.run(Thread.java:840) Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) 2024-12-05T15:14:35,260 WARN [Thread-815 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:35,265 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL 2024-12-05T15:14:35,265 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta with entries=11, filesize=3.63 KB; new WAL /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411675249.meta 2024-12-05T15:14:35,266 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:45209:45209),(127.0.0.1/127.0.0.1:43925:43925)] 2024-12-05T15:14:35,266 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta is not closed yet, will try archiving it next time 2024-12-05T15:14:35,266 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,266 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:43797,DS-5459db68-20a8-4270-b89c-e411daaa5d88,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:35,266 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta 2024-12-05T15:14:35,267 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd/.tmp/info/a0eb7b7e0376413cadfad1e5ac151c8c is 45, key is default/info:d/1733411628311/Put/seqid=0 2024-12-05T15:14:35,267 WARN [IPC Server handler 1 on default port 32987 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta has not been closed. Lease recovery is in progress. RecoveryId = 1051 for block blk_1073741834_1018 2024-12-05T15:14:35,267 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta after 1ms 2024-12-05T15:14:35,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741867_1052 (size=5037) 2024-12-05T15:14:35,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741867_1052 (size=5037) 2024-12-05T15:14:35,272 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd/.tmp/info/a0eb7b7e0376413cadfad1e5ac151c8c 2024-12-05T15:14:35,279 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd/.tmp/info/a0eb7b7e0376413cadfad1e5ac151c8c as hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd/info/a0eb7b7e0376413cadfad1e5ac151c8c 2024-12-05T15:14:35,285 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd/info/a0eb7b7e0376413cadfad1e5ac151c8c, entries=2, sequenceid=6, filesize=4.9 K 2024-12-05T15:14:35,286 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 474a835416f2f984a3141295bfac95bd in 38ms, sequenceid=6, compaction requested=false 2024-12-05T15:14:35,289 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/data/hbase/namespace/474a835416f2f984a3141295bfac95bd/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-05T15:14:35,290 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:14:35,290 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 474a835416f2f984a3141295bfac95bd: 2024-12-05T15:14:35,290 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733411627787.474a835416f2f984a3141295bfac95bd. 2024-12-05T15:14:35,291 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 817a20dfbf0d1531445f95325bcaf27c, disabling compactions & flushes 2024-12-05T15:14:35,291 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,291 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,291 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. after waiting 0 ms 2024-12-05T15:14:35,291 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,291 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 817a20dfbf0d1531445f95325bcaf27c: 2024-12-05T15:14:35,291 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:14:35,338 INFO [RS:1;4ecb0f57b706:39169 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,39169,1733411628416; zookeeper connection closed. 2024-12-05T15:14:35,338 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39169-0x10062be18070003, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:14:35,338 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@216a6fd0 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@216a6fd0 2024-12-05T15:14:35,448 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:14:35,448 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(3579): Received CLOSE for 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:14:35,448 DEBUG [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 817a20dfbf0d1531445f95325bcaf27c 2024-12-05T15:14:35,448 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 817a20dfbf0d1531445f95325bcaf27c, disabling compactions & flushes 2024-12-05T15:14:35,448 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:14:35,449 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,449 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. after waiting 0 ms 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 817a20dfbf0d1531445f95325bcaf27c: 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing TestLogRolling-testLogRollOnDatanodeDeath,,1733411628527.817a20dfbf0d1531445f95325bcaf27c. 2024-12-05T15:14:35,449 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing hbase:meta,,1.1588230740 2024-12-05T15:14:35,614 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 to hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs/4ecb0f57b706%2C44787%2C1733411626859.1733411627293 2024-12-05T15:14:35,616 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411651250 to hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs/4ecb0f57b706%2C44787%2C1733411626859.1733411651250 2024-12-05T15:14:35,617 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655263 to hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs/4ecb0f57b706%2C44787%2C1733411626859.1733411655263 2024-12-05T15:14:35,618 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.1733411655284 to hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/oldWALs/4ecb0f57b706%2C44787%2C1733411626859.1733411655284 2024-12-05T15:14:35,649 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1624): We were exiting though online regions are not empty, because some regions failed closing 2024-12-05T15:14:35,649 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,44787,1733411626859; all regions closed. 2024-12-05T15:14:35,649 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859 2024-12-05T15:14:35,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741866_1050 (size=93) 2024-12-05T15:14:35,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741866_1050 (size=93) 2024-12-05T15:14:35,759 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@66d38368[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741837_1013 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:35,759 WARN [java.util.concurrent.ThreadPoolExecutor$Worker@5d6988cd[State = -1, empty queue] {}] datanode.DataNode$DataTransfer(3129): DatanodeRegistration(127.0.0.1:32933, datanodeUuid=aee4fce9-277a-43d3-9380-0e0a2c5bf5c9, infoPort=45209, infoSecurePort=0, ipcPort=42531, storageInfo=lv=-57;cid=testClusterID;nsid=1662731957;c=1733411625957):Failed to transfer BP-308116525-172.17.0.2-1733411625957:blk_1073741835_1011 to 127.0.0.1:40345 got java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataNode$DataTransfer.run(DataNode.java:3063) ~[hadoop-hdfs-3.4.1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:36,242 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-05T15:14:36,242 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-05T15:14:36,481 INFO [regionserver/4ecb0f57b706:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:14:37,151 INFO [regionserver/4ecb0f57b706:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:14:37,816 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741857_1040 (size=450) 2024-12-05T15:14:39,268 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=1 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta after 4002ms 2024-12-05T15:14:40,291 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,308 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,310 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,310 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,316 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,317 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,652 ERROR [WAL-Shutdown-0 {}] wal.FSHLog(508): We have waited 5 seconds but the close of writer(s) doesn't complete.Please check the status of underlying filesystem or increase the wait time by the config "hbase.wal.fshlog.wait.on.shutdown.seconds" 2024-12-05T15:14:40,653 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859 2024-12-05T15:14:40,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741862_1046 (size=13280) 2024-12-05T15:14:40,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741862_1046 (size=13280) 2024-12-05T15:14:40,656 DEBUG [RS:0;4ecb0f57b706:44787 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:14:40,656 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:14:40,657 INFO [RS:0;4ecb0f57b706:44787 {}] hbase.ChoreService(370): Chore service for: regionserver/4ecb0f57b706:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-05T15:14:40,657 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:14:40,657 INFO [RS:0;4ecb0f57b706:44787 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:44787 2024-12-05T15:14:40,659 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:14:40,659 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/4ecb0f57b706,44787,1733411626859 2024-12-05T15:14:40,660 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [4ecb0f57b706,44787,1733411626859] 2024-12-05T15:14:40,660 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 4ecb0f57b706,44787,1733411626859; numProcessing=2 2024-12-05T15:14:40,662 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/4ecb0f57b706,44787,1733411626859 already deleted, retry=false 2024-12-05T15:14:40,662 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 4ecb0f57b706,44787,1733411626859 expired; onlineServers=0 2024-12-05T15:14:40,662 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,38149,1733411626804' ***** 2024-12-05T15:14:40,662 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T15:14:40,662 DEBUG [M:0;4ecb0f57b706:38149 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4bc9f11d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:14:40,662 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,38149,1733411626804 2024-12-05T15:14:40,662 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,38149,1733411626804; all regions closed. 2024-12-05T15:14:40,662 DEBUG [M:0;4ecb0f57b706:38149 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:14:40,662 DEBUG [M:0;4ecb0f57b706:38149 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T15:14:40,662 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T15:14:40,662 DEBUG [M:0;4ecb0f57b706:38149 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T15:14:40,662 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411627068 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411627068,5,FailOnTimeoutGroup] 2024-12-05T15:14:40,662 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411627068 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411627068,5,FailOnTimeoutGroup] 2024-12-05T15:14:40,663 INFO [M:0;4ecb0f57b706:38149 {}] hbase.ChoreService(370): Chore service for: master/4ecb0f57b706:0 had [] on shutdown 2024-12-05T15:14:40,663 DEBUG [M:0;4ecb0f57b706:38149 {}] master.HMaster(1733): Stopping service threads 2024-12-05T15:14:40,663 INFO [M:0;4ecb0f57b706:38149 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T15:14:40,663 INFO [M:0;4ecb0f57b706:38149 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T15:14:40,663 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T15:14:40,664 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T15:14:40,664 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:40,664 DEBUG [M:0;4ecb0f57b706:38149 {}] zookeeper.ZKUtil(347): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T15:14:40,664 WARN [M:0;4ecb0f57b706:38149 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T15:14:40,664 INFO [M:0;4ecb0f57b706:38149 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-05T15:14:40,664 INFO [M:0;4ecb0f57b706:38149 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T15:14:40,664 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:14:40,664 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:14:40,664 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:40,664 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:40,664 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:14:40,664 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:40,664 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=40.09 KB heapSize=49.30 KB 2024-12-05T15:14:40,681 DEBUG [M:0;4ecb0f57b706:38149 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/8fd8680473064f32859bfac42b44fca6 is 82, key is hbase:meta,,1/info:regioninfo/1733411627759/Put/seqid=0 2024-12-05T15:14:40,683 WARN [Thread-832 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741868_1053 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createSocketForPipeline(DataStreamer.java:256) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1894) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:40,683 WARN [Thread-832 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741868_1053 in pipeline [DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK], DatanodeInfoWithStorage[127.0.0.1:32933,DS-6781d264-5b19-4912-a89b-9a138120ec6a,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:40,683 WARN [Thread-832 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741868_1053 2024-12-05T15:14:40,684 WARN [Thread-832 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:40,690 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741869_1054 (size=5672) 2024-12-05T15:14:40,690 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741869_1054 (size=5672) 2024-12-05T15:14:40,690 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/8fd8680473064f32859bfac42b44fca6 2024-12-05T15:14:40,712 DEBUG [M:0;4ecb0f57b706:38149 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ca60755bad854d399db7c80ee4f0cf9c is 775, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733411628916/Put/seqid=0 2024-12-05T15:14:40,715 WARN [Thread-839 {}] hdfs.DataStreamer(1959): Exception in createBlockOutputStream blk_1073741870_1055 java.io.IOException: Got error, status=ERROR, status message , ack with firstBadLink as 127.0.0.1:40345 at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:128) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:104) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.createBlockOutputStream(DataStreamer.java:1947) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForCreate(DataStreamer.java:1842) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:752) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:14:40,715 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1231261664_22 at /127.0.0.1:37518 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741870_1055] {}] datanode.DataXceiver(892): DataNode{data=FSDataset{dirpath='[/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5, /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6]'}, localName='127.0.0.1:35103', datanodeUuid='9e57a236-2e7d-457a-9a0b-d39386031171', xmitsInProgress=0}:Exception transferring block BP-308116525-172.17.0.2-1733411625957:blk_1073741870_1055 to mirror 127.0.0.1:40345 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:40,715 WARN [Thread-839 {}] hdfs.DataStreamer(1731): Error Recovery for BP-308116525-172.17.0.2-1733411625957:blk_1073741870_1055 in pipeline [DatanodeInfoWithStorage[127.0.0.1:35103,DS-e76c7383-9a81-4a02-b02a-16d6e2df3bcd,DISK], DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK]) is bad. 2024-12-05T15:14:40,715 WARN [Thread-839 {}] hdfs.DataStreamer(1850): Abandoning BP-308116525-172.17.0.2-1733411625957:blk_1073741870_1055 2024-12-05T15:14:40,715 WARN [DataXceiver for client DFSClient_NONMAPREDUCE_1231261664_22 at /127.0.0.1:37518 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741870_1055] {}] datanode.BlockReceiver(316): Block 1073741870 has not released the reserved bytes. Releasing 134217728 bytes as part of close. 2024-12-05T15:14:40,715 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_1231261664_22 at /127.0.0.1:37518 [Receiving block BP-308116525-172.17.0.2-1733411625957:blk_1073741870_1055] {}] datanode.DataXceiver(331): 127.0.0.1:35103:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:37518 dst: /127.0.0.1:35103 java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:946) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:205) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:614) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:577) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:807) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:14:40,716 WARN [Thread-839 {}] hdfs.DataStreamer(1857): Excluding datanode DatanodeInfoWithStorage[127.0.0.1:40345,DS-4bb84d25-8bd3-497d-8de4-ea708ef3a6dc,DISK] 2024-12-05T15:14:40,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741871_1056 (size=7466) 2024-12-05T15:14:40,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741871_1056 (size=7466) 2024-12-05T15:14:40,721 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=39.42 KB at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ca60755bad854d399db7c80ee4f0cf9c 2024-12-05T15:14:40,742 DEBUG [M:0;4ecb0f57b706:38149 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ccbae472a0234489b787dd07ce467fcc is 69, key is 4ecb0f57b706,39169,1733411628416/rs:state/1733411628457/Put/seqid=0 2024-12-05T15:14:40,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741872_1057 (size=5224) 2024-12-05T15:14:40,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741872_1057 (size=5224) 2024-12-05T15:14:40,747 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=130 B at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ccbae472a0234489b787dd07ce467fcc 2024-12-05T15:14:40,760 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:14:40,761 INFO [RS:0;4ecb0f57b706:44787 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,44787,1733411626859; zookeeper connection closed. 2024-12-05T15:14:40,761 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:44787-0x10062be18070001, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:14:40,761 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@18be5ba9 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@18be5ba9 2024-12-05T15:14:40,761 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 2 regionserver(s) complete 2024-12-05T15:14:40,769 DEBUG [M:0;4ecb0f57b706:38149 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/9d3fee17c4d3417bbfa8d41b234d7703 is 52, key is load_balancer_on/state:d/1733411628399/Put/seqid=0 2024-12-05T15:14:40,774 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741873_1058 (size=5056) 2024-12-05T15:14:40,774 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741873_1058 (size=5056) 2024-12-05T15:14:40,775 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=97 (bloomFilter=true), to=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/9d3fee17c4d3417bbfa8d41b234d7703 2024-12-05T15:14:40,781 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/8fd8680473064f32859bfac42b44fca6 as hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/8fd8680473064f32859bfac42b44fca6 2024-12-05T15:14:40,786 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/8fd8680473064f32859bfac42b44fca6, entries=8, sequenceid=97, filesize=5.5 K 2024-12-05T15:14:40,787 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ca60755bad854d399db7c80ee4f0cf9c as hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ca60755bad854d399db7c80ee4f0cf9c 2024-12-05T15:14:40,793 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ca60755bad854d399db7c80ee4f0cf9c, entries=11, sequenceid=97, filesize=7.3 K 2024-12-05T15:14:40,794 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/ccbae472a0234489b787dd07ce467fcc as hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ccbae472a0234489b787dd07ce467fcc 2024-12-05T15:14:40,799 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/ccbae472a0234489b787dd07ce467fcc, entries=2, sequenceid=97, filesize=5.1 K 2024-12-05T15:14:40,800 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/9d3fee17c4d3417bbfa8d41b234d7703 as hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/9d3fee17c4d3417bbfa8d41b234d7703 2024-12-05T15:14:40,805 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/9d3fee17c4d3417bbfa8d41b234d7703, entries=1, sequenceid=97, filesize=4.9 K 2024-12-05T15:14:40,806 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(3040): Finished flush of dataSize ~40.09 KB/41051, heapSize ~49.23 KB/50416, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 142ms, sequenceid=97, compaction requested=false 2024-12-05T15:14:40,808 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:40,808 DEBUG [M:0;4ecb0f57b706:38149 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:14:40,808 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804 2024-12-05T15:14:40,810 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:32933 is added to blk_1073741861_1044 (size=757) 2024-12-05T15:14:40,810 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:35103 is added to blk_1073741861_1044 (size=757) 2024-12-05T15:14:40,825 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:14:40,827 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,840 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,841 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,841 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,845 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:40,845 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:41,269 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:42,270 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:43,270 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:44,271 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:44,991 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-05T15:14:44,991 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-05T15:14:45,272 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:45,811 ERROR [WAL-Shutdown-0 {}] wal.FSHLog(508): We have waited 5 seconds but the close of writer(s) doesn't complete.Please check the status of underlying filesystem or increase the wait time by the config "hbase.wal.fshlog.wait.on.shutdown.seconds" 2024-12-05T15:14:45,811 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:14:45,811 INFO [M:0;4ecb0f57b706:38149 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-05T15:14:45,811 INFO [M:0;4ecb0f57b706:38149 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:38149 2024-12-05T15:14:45,813 DEBUG [M:0;4ecb0f57b706:38149 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/4ecb0f57b706,38149,1733411626804 already deleted, retry=false 2024-12-05T15:14:45,915 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:14:45,915 INFO [M:0;4ecb0f57b706:38149 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,38149,1733411626804; zookeeper connection closed. 2024-12-05T15:14:45,915 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:38149-0x10062be18070000, quorum=127.0.0.1:55344, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:14:45,918 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@7d38b2e7{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:14:45,918 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1783a9ee{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:14:45,918 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:14:45,919 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@601bb964{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:14:45,919 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@20e478b8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,STOPPED} 2024-12-05T15:14:45,920 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:14:45,920 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:14:45,920 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-308116525-172.17.0.2-1733411625957 (Datanode Uuid aee4fce9-277a-43d3-9380-0e0a2c5bf5c9) service to localhost/127.0.0.1:32987 2024-12-05T15:14:45,920 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:14:45,921 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data3/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:45,921 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data4/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:45,921 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:14:45,926 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@75f9f533{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:14:45,926 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@154ff60c{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:14:45,926 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:14:45,927 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6e33bbd1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:14:45,927 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3a5b3132{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,STOPPED} 2024-12-05T15:14:45,928 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:14:45,928 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:14:45,928 WARN [BP-308116525-172.17.0.2-1733411625957 heartbeating to localhost/127.0.0.1:32987 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-308116525-172.17.0.2-1733411625957 (Datanode Uuid 9e57a236-2e7d-457a-9a0b-d39386031171) service to localhost/127.0.0.1:32987 2024-12-05T15:14:45,928 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:14:45,928 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data5/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:45,929 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/cluster_fdaf04df-170f-4287-89c7-230b8e59ca95/dfs/data/data6/current/BP-308116525-172.17.0.2-1733411625957 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:14:45,929 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:14:45,935 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@55bfb256{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:14:45,936 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@7e1ef301{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:14:45,936 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:14:45,936 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2c0370f2{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:14:45,936 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1101772a{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir/,STOPPED} 2024-12-05T15:14:45,944 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-05T15:14:45,972 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-05T15:14:45,982 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnDatanodeDeath Thread=86 (was 66) Potentially hanging thread: nioEventLoopGroup-17-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-6-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-11-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:32987 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Timer for 'DataNode' metrics system java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: Abort regionserver monitor java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: nioEventLoopGroup-10-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-5-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-16-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-3 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-10-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-17-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-16-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:32987 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-5-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-11-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:32987 from jenkins.hfs.1 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: RS-EventLoopGroup-5-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-10-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:32987 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Close-WAL-Writer-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:175) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$$Lambda$798/0x00007eff88b8d7e8.run(Unknown Source) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-6-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:32987 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: Close-WAL-Writer-0 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:175) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) app//org.apache.hadoop.hbase.regionserver.wal.FSHLog$$Lambda$798/0x00007eff88b8d7e8.run(Unknown Source) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:32987 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-11-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-4 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-6-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-17-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-7-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-16-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.2@localhost:32987 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:32987 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-7-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-7-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.1@localhost:32987 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=414 (was 401) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=168 (was 226), ProcessCount=11 (was 11), AvailableMemoryMB=8674 (was 9072) 2024-12-05T15:14:45,989 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnPipelineRestart Thread=86, OpenFileDescriptor=414, MaxFileDescriptor=1048576, SystemLoadAverage=168, ProcessCount=11, AvailableMemoryMB=8674 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.log.dir so I do NOT create it in target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/9e1483d4-8ae8-46a1-6f6d-3de2821fa40f/hadoop.tmp.dir so I do NOT create it in target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c, deleteOnExit=true 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/test.cache.data in system properties and HBase conf 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T15:14:45,990 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-05T15:14:45,991 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T15:14:45,991 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:14:45,992 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:14:45,992 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T15:14:45,992 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/nfs.dump.dir in system properties and HBase conf 2024-12-05T15:14:45,992 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir in system properties and HBase conf 2024-12-05T15:14:45,992 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:14:45,992 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T15:14:45,992 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T15:14:46,005 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:14:46,078 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:14:46,082 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:14:46,088 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:14:46,088 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:14:46,088 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:14:46,089 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:14:46,089 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@64c90eeb{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:14:46,090 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@134868e9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:14:46,098 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:46,204 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@29c97027{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir/jetty-localhost-42665-hadoop-hdfs-3_4_1-tests_jar-_-any-16723562578621384128/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:14:46,205 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@38cff363{HTTP/1.1, (http/1.1)}{localhost:42665} 2024-12-05T15:14:46,205 INFO [Time-limited test {}] server.Server(415): Started @177020ms 2024-12-05T15:14:46,219 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:14:46,273 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:46,289 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:14:46,292 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:14:46,293 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:14:46,293 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:14:46,293 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:14:46,293 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@79c9ffa9{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:14:46,294 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@48245567{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:14:46,416 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@71404fe1{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir/jetty-localhost-37515-hadoop-hdfs-3_4_1-tests_jar-_-any-8090836932445588610/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:14:46,417 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@48bfabaa{HTTP/1.1, (http/1.1)}{localhost:37515} 2024-12-05T15:14:46,417 INFO [Time-limited test {}] server.Server(415): Started @177233ms 2024-12-05T15:14:46,418 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:14:46,451 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:14:46,455 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:14:46,456 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:14:46,456 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:14:46,456 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:14:46,457 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@54840b6b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:14:46,458 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@775e1fb9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:14:46,524 WARN [Thread-933 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data1/current/BP-107838756-172.17.0.2-1733411686024/current, will proceed with Du for space computation calculation, 2024-12-05T15:14:46,524 WARN [Thread-934 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data2/current/BP-107838756-172.17.0.2-1733411686024/current, will proceed with Du for space computation calculation, 2024-12-05T15:14:46,547 WARN [Thread-912 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:14:46,550 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xbd2cf0c46372bf7 with lease ID 0xf7e183916458b2fa: Processing first storage report for DS-c04dc19c-ae7f-4dc6-a523-21295a18894d from datanode DatanodeRegistration(127.0.0.1:37965, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=35827, infoSecurePort=0, ipcPort=42783, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024) 2024-12-05T15:14:46,551 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbd2cf0c46372bf7 with lease ID 0xf7e183916458b2fa: from storage DS-c04dc19c-ae7f-4dc6-a523-21295a18894d node DatanodeRegistration(127.0.0.1:37965, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=35827, infoSecurePort=0, ipcPort=42783, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:14:46,551 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xbd2cf0c46372bf7 with lease ID 0xf7e183916458b2fa: Processing first storage report for DS-1208898a-81c7-4643-a302-3b3a089c1dd8 from datanode DatanodeRegistration(127.0.0.1:37965, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=35827, infoSecurePort=0, ipcPort=42783, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024) 2024-12-05T15:14:46,551 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbd2cf0c46372bf7 with lease ID 0xf7e183916458b2fa: from storage DS-1208898a-81c7-4643-a302-3b3a089c1dd8 node DatanodeRegistration(127.0.0.1:37965, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=35827, infoSecurePort=0, ipcPort=42783, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:14:46,580 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@54926efb{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir/jetty-localhost-39203-hadoop-hdfs-3_4_1-tests_jar-_-any-13924714937652526038/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:14:46,580 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@31e65952{HTTP/1.1, (http/1.1)}{localhost:39203} 2024-12-05T15:14:46,580 INFO [Time-limited test {}] server.Server(415): Started @177396ms 2024-12-05T15:14:46,581 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:14:46,677 WARN [Thread-959 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data3/current/BP-107838756-172.17.0.2-1733411686024/current, will proceed with Du for space computation calculation, 2024-12-05T15:14:46,678 WARN [Thread-960 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data4/current/BP-107838756-172.17.0.2-1733411686024/current, will proceed with Du for space computation calculation, 2024-12-05T15:14:46,696 WARN [Thread-948 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:14:46,698 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x785c953ba05e49bf with lease ID 0xf7e183916458b2fb: Processing first storage report for DS-3579be7c-0bfd-45c7-aa32-85fca803b340 from datanode DatanodeRegistration(127.0.0.1:46547, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=40309, infoSecurePort=0, ipcPort=36675, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024) 2024-12-05T15:14:46,698 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x785c953ba05e49bf with lease ID 0xf7e183916458b2fb: from storage DS-3579be7c-0bfd-45c7-aa32-85fca803b340 node DatanodeRegistration(127.0.0.1:46547, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=40309, infoSecurePort=0, ipcPort=36675, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:14:46,699 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x785c953ba05e49bf with lease ID 0xf7e183916458b2fb: Processing first storage report for DS-736508bc-8897-4ab4-a855-8f92fe91d984 from datanode DatanodeRegistration(127.0.0.1:46547, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=40309, infoSecurePort=0, ipcPort=36675, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024) 2024-12-05T15:14:46,699 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x785c953ba05e49bf with lease ID 0xf7e183916458b2fb: from storage DS-736508bc-8897-4ab4-a855-8f92fe91d984 node DatanodeRegistration(127.0.0.1:46547, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=40309, infoSecurePort=0, ipcPort=36675, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:14:46,704 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18 2024-12-05T15:14:46,707 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/zookeeper_0, clientPort=50082, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T15:14:46,708 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=50082 2024-12-05T15:14:46,708 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:46,710 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:46,718 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:14:46,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:14:46,720 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b with version=8 2024-12-05T15:14:46,720 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/hbase-staging 2024-12-05T15:14:46,722 INFO [Time-limited test {}] client.ConnectionUtils(129): master/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:14:46,722 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:14:46,722 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:14:46,722 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:14:46,722 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:14:46,722 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:14:46,722 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:14:46,723 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:14:46,723 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:34185 2024-12-05T15:14:46,724 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:46,725 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:46,727 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:34185 connecting to ZooKeeper ensemble=127.0.0.1:50082 2024-12-05T15:14:46,735 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:341850x0, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:14:46,735 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:34185-0x10062bf02190000 connected 2024-12-05T15:14:46,758 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:14:46,758 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:14:46,759 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:14:46,759 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=34185 2024-12-05T15:14:46,759 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=34185 2024-12-05T15:14:46,760 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=34185 2024-12-05T15:14:46,760 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=34185 2024-12-05T15:14:46,760 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=34185 2024-12-05T15:14:46,760 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b, hbase.cluster.distributed=false 2024-12-05T15:14:46,777 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:14:46,777 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:14:46,777 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:14:46,777 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:14:46,777 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:14:46,777 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:14:46,777 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:14:46,777 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:14:46,778 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:32857 2024-12-05T15:14:46,778 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T15:14:46,779 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T15:14:46,780 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:46,781 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:46,785 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:32857 connecting to ZooKeeper ensemble=127.0.0.1:50082 2024-12-05T15:14:46,788 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:328570x0, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:14:46,788 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:328570x0, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:14:46,788 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:32857-0x10062bf02190001 connected 2024-12-05T15:14:46,789 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:14:46,789 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:14:46,789 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=32857 2024-12-05T15:14:46,790 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=32857 2024-12-05T15:14:46,790 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=32857 2024-12-05T15:14:46,790 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=32857 2024-12-05T15:14:46,790 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=32857 2024-12-05T15:14:46,791 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:46,793 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:14:46,793 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:14:46,793 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:46,795 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:14:46,795 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,795 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:14:46,795 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,796 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:14:46,796 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/4ecb0f57b706,34185,1733411686721 from backup master directory 2024-12-05T15:14:46,798 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:46,798 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:14:46,798 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:14:46,798 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:14:46,798 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:14:46,798 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:46,804 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;4ecb0f57b706:34185 2024-12-05T15:14:46,811 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:14:46,811 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:14:46,812 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/hbase.id with ID: 805d9ec4-f27f-4056-8aed-934b1a0b0793 2024-12-05T15:14:46,823 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:46,826 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,826 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:14:46,833 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:14:46,834 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:14:46,834 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T15:14:46,834 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:14:46,841 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:14:46,842 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:14:46,843 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store 2024-12-05T15:14:46,848 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:14:46,849 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:14:46,849 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:46,849 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:14:46,849 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:46,849 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:46,849 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:14:46,849 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:46,849 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:14:46,849 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:14:46,850 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/.initializing 2024-12-05T15:14:46,850 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:46,853 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C34185%2C1733411686721, suffix=, logDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721, archiveDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/oldWALs, maxLogs=10 2024-12-05T15:14:46,854 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C34185%2C1733411686721.1733411686854 2024-12-05T15:14:46,859 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 2024-12-05T15:14:46,859 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35827:35827),(127.0.0.1/127.0.0.1:40309:40309)] 2024-12-05T15:14:46,859 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:14:46,860 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:46,860 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,860 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,861 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,863 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T15:14:46,863 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,863 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:14:46,863 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,864 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T15:14:46,865 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,865 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:14:46,865 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,866 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T15:14:46,866 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,867 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:14:46,867 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,868 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T15:14:46,868 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,869 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:14:46,869 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,870 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,871 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T15:14:46,873 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:14:46,875 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:14:46,875 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=782797, jitterRate=-0.004622265696525574}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T15:14:46,876 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:14:46,876 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T15:14:46,880 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7694c5d0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:14:46,881 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-05T15:14:46,881 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T15:14:46,881 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T15:14:46,881 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T15:14:46,882 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-05T15:14:46,882 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-05T15:14:46,882 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T15:14:46,884 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T15:14:46,885 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T15:14:46,886 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-05T15:14:46,886 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T15:14:46,887 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T15:14:46,888 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-05T15:14:46,888 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T15:14:46,889 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T15:14:46,891 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-05T15:14:46,892 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T15:14:46,893 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T15:14:46,894 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T15:14:46,896 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T15:14:46,898 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:14:46,898 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:14:46,898 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,898 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,898 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=4ecb0f57b706,34185,1733411686721, sessionid=0x10062bf02190000, setting cluster-up flag (Was=false) 2024-12-05T15:14:46,901 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,901 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,907 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T15:14:46,908 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:46,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:46,916 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T15:14:46,917 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:46,920 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-05T15:14:46,920 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-05T15:14:46,920 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T15:14:46,920 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 4ecb0f57b706,34185,1733411686721 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/4ecb0f57b706:0, corePoolSize=10, maxPoolSize=10 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:14:46,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:46,922 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733411716922 2024-12-05T15:14:46,922 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T15:14:46,922 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T15:14:46,922 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T15:14:46,922 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T15:14:46,922 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T15:14:46,922 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T15:14:46,923 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:46,923 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:14:46,923 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-05T15:14:46,923 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T15:14:46,923 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T15:14:46,923 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T15:14:46,923 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T15:14:46,924 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T15:14:46,924 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411686924,5,FailOnTimeoutGroup] 2024-12-05T15:14:46,924 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,924 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411686924,5,FailOnTimeoutGroup] 2024-12-05T15:14:46,924 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:46,924 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T15:14:46,924 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:46,924 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:46,924 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:14:46,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:14:46,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:14:46,931 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-05T15:14:46,931 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b 2024-12-05T15:14:46,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:14:46,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:14:46,940 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:46,941 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:14:46,942 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:14:46,943 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,943 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:14:46,943 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:14:46,944 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:14:46,945 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,945 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:14:46,945 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:14:46,946 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:14:46,946 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:46,947 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:14:46,947 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/meta/1588230740 2024-12-05T15:14:46,948 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/meta/1588230740 2024-12-05T15:14:46,949 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:14:46,950 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:14:46,952 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:14:46,953 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=848897, jitterRate=0.07942868769168854}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:14:46,953 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:14:46,953 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:14:46,953 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:14:46,953 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:14:46,953 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:14:46,953 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:14:46,954 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:14:46,954 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:14:46,955 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:14:46,955 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-05T15:14:46,955 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T15:14:46,956 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T15:14:46,957 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T15:14:47,003 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;4ecb0f57b706:32857 2024-12-05T15:14:47,004 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1008): ClusterId : 805d9ec4-f27f-4056-8aed-934b1a0b0793 2024-12-05T15:14:47,004 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T15:14:47,006 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T15:14:47,006 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T15:14:47,008 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T15:14:47,009 DEBUG [RS:0;4ecb0f57b706:32857 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@38aee027, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:14:47,009 DEBUG [RS:0;4ecb0f57b706:32857 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@56b0ef89, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:14:47,009 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-05T15:14:47,009 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-05T15:14:47,010 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-05T15:14:47,010 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,34185,1733411686721 with isa=4ecb0f57b706/172.17.0.2:32857, startcode=1733411686776 2024-12-05T15:14:47,010 DEBUG [RS:0;4ecb0f57b706:32857 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:14:47,012 INFO [RS-EventLoopGroup-8-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:36069, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:14:47,013 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34185 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,013 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=34185 {}] master.ServerManager(486): Registering regionserver=4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,015 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b 2024-12-05T15:14:47,015 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:33521 2024-12-05T15:14:47,015 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-05T15:14:47,017 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:14:47,017 DEBUG [RS:0;4ecb0f57b706:32857 {}] zookeeper.ZKUtil(111): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,017 WARN [RS:0;4ecb0f57b706:32857 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:14:47,017 INFO [RS:0;4ecb0f57b706:32857 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:14:47,017 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,017 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [4ecb0f57b706,32857,1733411686776] 2024-12-05T15:14:47,020 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-05T15:14:47,020 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T15:14:47,022 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T15:14:47,022 INFO [RS:0;4ecb0f57b706:32857 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:14:47,022 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,023 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-05T15:14:47,023 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:14:47,024 DEBUG [RS:0;4ecb0f57b706:32857 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:14:47,025 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,025 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,025 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,025 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,025 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,32857,1733411686776-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:14:47,042 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T15:14:47,042 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,32857,1733411686776-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,057 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.Replication(204): 4ecb0f57b706,32857,1733411686776 started 2024-12-05T15:14:47,057 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1767): Serving as 4ecb0f57b706,32857,1733411686776, RpcServer on 4ecb0f57b706/172.17.0.2:32857, sessionid=0x10062bf02190001 2024-12-05T15:14:47,057 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T15:14:47,057 DEBUG [RS:0;4ecb0f57b706:32857 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,057 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,32857,1733411686776' 2024-12-05T15:14:47,057 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T15:14:47,057 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T15:14:47,058 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T15:14:47,058 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T15:14:47,058 DEBUG [RS:0;4ecb0f57b706:32857 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,058 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,32857,1733411686776' 2024-12-05T15:14:47,058 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T15:14:47,058 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T15:14:47,059 DEBUG [RS:0;4ecb0f57b706:32857 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T15:14:47,059 INFO [RS:0;4ecb0f57b706:32857 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T15:14:47,059 INFO [RS:0;4ecb0f57b706:32857 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T15:14:47,098 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:47,107 WARN [4ecb0f57b706:34185 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-05T15:14:47,161 INFO [RS:0;4ecb0f57b706:32857 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C32857%2C1733411686776, suffix=, logDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776, archiveDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/oldWALs, maxLogs=32 2024-12-05T15:14:47,162 INFO [RS:0;4ecb0f57b706:32857 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:14:47,169 INFO [RS:0;4ecb0f57b706:32857 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:14:47,169 DEBUG [RS:0;4ecb0f57b706:32857 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:40309:40309),(127.0.0.1/127.0.0.1:35827:35827)] 2024-12-05T15:14:47,274 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:47,358 DEBUG [4ecb0f57b706:34185 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-05T15:14:47,358 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,359 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,32857,1733411686776, state=OPENING 2024-12-05T15:14:47,361 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T15:14:47,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:47,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:47,363 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=4ecb0f57b706,32857,1733411686776}] 2024-12-05T15:14:47,363 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:14:47,363 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:14:47,516 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,517 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T15:14:47,519 INFO [RS-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:44994, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T15:14:47,522 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-05T15:14:47,522 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:14:47,524 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C32857%2C1733411686776.meta, suffix=.meta, logDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776, archiveDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/oldWALs, maxLogs=32 2024-12-05T15:14:47,525 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta 2024-12-05T15:14:47,533 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:35827:35827),(127.0.0.1/127.0.0.1:40309:40309)] 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T15:14:47,534 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-05T15:14:47,534 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-05T15:14:47,536 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:14:47,537 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:14:47,537 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:47,537 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:14:47,538 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:14:47,538 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:14:47,539 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:47,539 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:14:47,539 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:14:47,540 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:14:47,540 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:47,540 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:14:47,541 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/meta/1588230740 2024-12-05T15:14:47,542 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/meta/1588230740 2024-12-05T15:14:47,544 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:14:47,545 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:14:47,546 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=784899, jitterRate=-0.0019500106573104858}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:14:47,547 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:14:47,548 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733411687516 2024-12-05T15:14:47,550 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T15:14:47,550 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-05T15:14:47,551 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,552 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,32857,1733411686776, state=OPEN 2024-12-05T15:14:47,557 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:14:47,557 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:14:47,557 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:14:47,557 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:14:47,559 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T15:14:47,559 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=4ecb0f57b706,32857,1733411686776 in 194 msec 2024-12-05T15:14:47,561 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T15:14:47,561 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 604 msec 2024-12-05T15:14:47,563 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 644 msec 2024-12-05T15:14:47,563 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733411687563, completionTime=-1 2024-12-05T15:14:47,563 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-05T15:14:47,564 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-05T15:14:47,564 DEBUG [hconnection-0x5749a0ee-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:14:47,566 INFO [RS-EventLoopGroup-9-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45010, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:14:47,567 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-05T15:14:47,567 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733411747567 2024-12-05T15:14:47,567 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733411807567 2024-12-05T15:14:47,567 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 3 msec 2024-12-05T15:14:47,573 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,34185,1733411686721-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,573 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,34185,1733411686721-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,573 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,34185,1733411686721-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,573 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-4ecb0f57b706:34185, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,573 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:47,573 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-05T15:14:47,573 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:14:47,574 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-05T15:14:47,574 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-05T15:14:47,575 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:14:47,576 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:47,576 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:14:47,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:14:47,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:14:47,585 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 9a5acd1c15e040330ff7ee9cb3eb52c2, NAME => 'hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b 2024-12-05T15:14:47,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:14:47,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:14:47,592 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:47,592 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 9a5acd1c15e040330ff7ee9cb3eb52c2, disabling compactions & flushes 2024-12-05T15:14:47,592 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:14:47,592 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:14:47,592 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. after waiting 0 ms 2024-12-05T15:14:47,592 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:14:47,592 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:14:47,592 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 9a5acd1c15e040330ff7ee9cb3eb52c2: 2024-12-05T15:14:47,594 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:14:47,594 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733411687594"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411687594"}]},"ts":"1733411687594"} 2024-12-05T15:14:47,596 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:14:47,597 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:14:47,597 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411687597"}]},"ts":"1733411687597"} 2024-12-05T15:14:47,599 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-05T15:14:47,603 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=9a5acd1c15e040330ff7ee9cb3eb52c2, ASSIGN}] 2024-12-05T15:14:47,604 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=9a5acd1c15e040330ff7ee9cb3eb52c2, ASSIGN 2024-12-05T15:14:47,605 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=9a5acd1c15e040330ff7ee9cb3eb52c2, ASSIGN; state=OFFLINE, location=4ecb0f57b706,32857,1733411686776; forceNewPlan=false, retain=false 2024-12-05T15:14:47,756 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=9a5acd1c15e040330ff7ee9cb3eb52c2, regionState=OPENING, regionLocation=4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,758 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 9a5acd1c15e040330ff7ee9cb3eb52c2, server=4ecb0f57b706,32857,1733411686776}] 2024-12-05T15:14:47,911 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,915 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:14:47,915 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 9a5acd1c15e040330ff7ee9cb3eb52c2, NAME => 'hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:14:47,916 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:14:47,916 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:47,916 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:14:47,916 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:14:47,917 INFO [StoreOpener-9a5acd1c15e040330ff7ee9cb3eb52c2-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:14:47,919 INFO [StoreOpener-9a5acd1c15e040330ff7ee9cb3eb52c2-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 9a5acd1c15e040330ff7ee9cb3eb52c2 columnFamilyName info 2024-12-05T15:14:47,919 DEBUG [StoreOpener-9a5acd1c15e040330ff7ee9cb3eb52c2-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:47,919 INFO [StoreOpener-9a5acd1c15e040330ff7ee9cb3eb52c2-1 {}] regionserver.HStore(327): Store=9a5acd1c15e040330ff7ee9cb3eb52c2/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:14:47,920 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/namespace/9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:14:47,921 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/namespace/9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:14:47,922 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:14:47,925 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/hbase/namespace/9a5acd1c15e040330ff7ee9cb3eb52c2/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:14:47,925 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 9a5acd1c15e040330ff7ee9cb3eb52c2; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=743395, jitterRate=-0.05472448468208313}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:14:47,926 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 9a5acd1c15e040330ff7ee9cb3eb52c2: 2024-12-05T15:14:47,927 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2., pid=6, masterSystemTime=1733411687911 2024-12-05T15:14:47,929 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:14:47,929 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:14:47,929 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=9a5acd1c15e040330ff7ee9cb3eb52c2, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:47,934 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T15:14:47,934 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 9a5acd1c15e040330ff7ee9cb3eb52c2, server=4ecb0f57b706,32857,1733411686776 in 173 msec 2024-12-05T15:14:47,936 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T15:14:47,936 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=9a5acd1c15e040330ff7ee9cb3eb52c2, ASSIGN in 331 msec 2024-12-05T15:14:47,937 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:14:47,937 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411687937"}]},"ts":"1733411687937"} 2024-12-05T15:14:47,939 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-05T15:14:47,941 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:14:47,943 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 368 msec 2024-12-05T15:14:47,976 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-05T15:14:47,978 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:14:47,978 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:47,978 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:14:47,983 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-05T15:14:47,991 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:14:47,994 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 11 msec 2024-12-05T15:14:48,005 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-05T15:14:48,012 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:14:48,015 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 10 msec 2024-12-05T15:14:48,030 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-05T15:14:48,033 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-05T15:14:48,033 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.235sec 2024-12-05T15:14:48,034 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T15:14:48,034 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T15:14:48,034 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T15:14:48,034 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T15:14:48,034 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T15:14:48,034 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,34185,1733411686721-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:14:48,034 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,34185,1733411686721-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T15:14:48,036 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-05T15:14:48,036 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T15:14:48,036 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,34185,1733411686721-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:14:48,093 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1ed3dd55 to 127.0.0.1:50082 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@44061808 2024-12-05T15:14:48,097 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@498a3293, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:14:48,098 DEBUG [hconnection-0x3d38b5f2-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:14:48,099 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:48,101 INFO [RS-EventLoopGroup-9-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45020, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:14:48,103 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=4ecb0f57b706,34185,1733411686721 2024-12-05T15:14:48,103 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:14:48,107 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-05T15:14:48,107 INFO [Time-limited test {}] wal.TestLogRolling(297): Starting testLogRollOnPipelineRestart 2024-12-05T15:14:48,107 INFO [Time-limited test {}] wal.TestLogRolling(300): Replication=2 2024-12-05T15:14:48,107 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-05T15:14:48,109 INFO [RS-EventLoopGroup-8-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:46568, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-05T15:14:48,110 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34185 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-05T15:14:48,111 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34185 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-05T15:14:48,111 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34185 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testLogRollOnPipelineRestart', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:14:48,112 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34185 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart 2024-12-05T15:14:48,113 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:14:48,113 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:48,113 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34185 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRollOnPipelineRestart" procId is: 9 2024-12-05T15:14:48,114 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34185 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:14:48,114 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:14:48,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741837_1013 (size=395) 2024-12-05T15:14:48,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741837_1013 (size=395) 2024-12-05T15:14:48,123 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 63c9f1b847ea3792bcfec0f2259fa595, NAME => 'TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRollOnPipelineRestart', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b 2024-12-05T15:14:48,130 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37965 is added to blk_1073741838_1014 (size=78) 2024-12-05T15:14:48,130 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46547 is added to blk_1073741838_1014 (size=78) 2024-12-05T15:14:48,130 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:48,130 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1681): Closing 63c9f1b847ea3792bcfec0f2259fa595, disabling compactions & flushes 2024-12-05T15:14:48,130 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:48,130 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:48,131 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. after waiting 1 ms 2024-12-05T15:14:48,131 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:48,131 INFO [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:48,131 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRollOnPipelineRestart-pool-0 {}] regionserver.HRegion(1635): Region close journal for 63c9f1b847ea3792bcfec0f2259fa595: 2024-12-05T15:14:48,132 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:14:48,132 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595.","families":{"info":[{"qualifier":"regioninfo","vlen":77,"tag":[],"timestamp":"1733411688132"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411688132"}]},"ts":"1733411688132"} 2024-12-05T15:14:48,134 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:14:48,135 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:14:48,135 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnPipelineRestart","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411688135"}]},"ts":"1733411688135"} 2024-12-05T15:14:48,137 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnPipelineRestart, state=ENABLING in hbase:meta 2024-12-05T15:14:48,141 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=63c9f1b847ea3792bcfec0f2259fa595, ASSIGN}] 2024-12-05T15:14:48,142 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=63c9f1b847ea3792bcfec0f2259fa595, ASSIGN 2024-12-05T15:14:48,143 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=63c9f1b847ea3792bcfec0f2259fa595, ASSIGN; state=OFFLINE, location=4ecb0f57b706,32857,1733411686776; forceNewPlan=false, retain=false 2024-12-05T15:14:48,274 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:48,294 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=63c9f1b847ea3792bcfec0f2259fa595, regionState=OPENING, regionLocation=4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:48,296 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 63c9f1b847ea3792bcfec0f2259fa595, server=4ecb0f57b706,32857,1733411686776}] 2024-12-05T15:14:48,449 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:48,453 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:48,453 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 63c9f1b847ea3792bcfec0f2259fa595, NAME => 'TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:14:48,453 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRollOnPipelineRestart 63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:14:48,453 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:14:48,453 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:14:48,453 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:14:48,455 INFO [StoreOpener-63c9f1b847ea3792bcfec0f2259fa595-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:14:48,456 INFO [StoreOpener-63c9f1b847ea3792bcfec0f2259fa595-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 63c9f1b847ea3792bcfec0f2259fa595 columnFamilyName info 2024-12-05T15:14:48,456 DEBUG [StoreOpener-63c9f1b847ea3792bcfec0f2259fa595-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:14:48,457 INFO [StoreOpener-63c9f1b847ea3792bcfec0f2259fa595-1 {}] regionserver.HStore(327): Store=63c9f1b847ea3792bcfec0f2259fa595/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:14:48,458 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:14:48,458 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:14:48,460 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:14:48,462 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:14:48,463 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened 63c9f1b847ea3792bcfec0f2259fa595; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=738626, jitterRate=-0.06078958511352539}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:14:48,464 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 63c9f1b847ea3792bcfec0f2259fa595: 2024-12-05T15:14:48,465 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595., pid=11, masterSystemTime=1733411688448 2024-12-05T15:14:48,467 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:48,467 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:48,468 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=63c9f1b847ea3792bcfec0f2259fa595, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,32857,1733411686776 2024-12-05T15:14:48,472 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-05T15:14:48,472 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 63c9f1b847ea3792bcfec0f2259fa595, server=4ecb0f57b706,32857,1733411686776 in 174 msec 2024-12-05T15:14:48,474 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-05T15:14:48,474 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRollOnPipelineRestart, region=63c9f1b847ea3792bcfec0f2259fa595, ASSIGN in 331 msec 2024-12-05T15:14:48,475 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:14:48,475 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRollOnPipelineRestart","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411688475"}]},"ts":"1733411688475"} 2024-12-05T15:14:48,477 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRollOnPipelineRestart, state=ENABLED in hbase:meta 2024-12-05T15:14:48,480 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:14:48,481 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testLogRollOnPipelineRestart in 369 msec 2024-12-05T15:14:49,100 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:49,275 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:50,100 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:50,276 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:50,494 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:14:50,506 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:50,507 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:50,507 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:50,507 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:50,507 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:50,508 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:50,511 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:50,513 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:14:51,101 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:51,277 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:52,102 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:52,278 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:53,021 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-05T15:14:53,021 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRollOnPipelineRestart' 2024-12-05T15:14:53,102 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:53,279 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:54,103 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:54,279 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:54,991 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2024-12-05T15:14:54,991 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2024-12-05T15:14:54,991 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart 2024-12-05T15:14:54,991 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart Metrics about Tables on a single HBase RegionServer 2024-12-05T15:14:54,992 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2024-12-05T15:14:54,992 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2024-12-05T15:14:55,104 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:55,280 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:56,105 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:56,281 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:57,105 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:57,281 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:58,106 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:58,116 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=34185 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:14:58,116 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testLogRollOnPipelineRestart, procId: 9 completed 2024-12-05T15:14:58,119 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testLogRollOnPipelineRestart 2024-12-05T15:14:58,119 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:14:58,282 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:59,106 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:14:59,282 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:00,107 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:00,125 INFO [Time-limited test {}] wal.TestLogRolling(337): log.getCurrentFileName()): hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:15:00,126 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1009 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1009 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:00,126 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006 java.io.IOException: Bad response ERROR for BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006 from datanode DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:00,126 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010 java.io.IOException: Bad response ERROR for BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010 from datanode DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:00,126 WARN [DataStreamer for file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 block BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1009 {}] hdfs.DataStreamer(1731): Error Recovery for BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1009 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK], DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]) is bad. 2024-12-05T15:15:00,126 WARN [DataStreamer for file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta block BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010 {}] hdfs.DataStreamer(1731): Error Recovery for BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010 in pipeline [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK], DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]) is bad. 2024-12-05T15:15:00,127 WARN [DataStreamer for file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 block BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006 {}] hdfs.DataStreamer(1731): Error Recovery for BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006 in pipeline [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK], DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:46547,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]) is bad. 2024-12-05T15:15:00,127 WARN [PacketResponder: BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:46547] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,127 WARN [PacketResponder: BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:46547] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Broken pipe at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,127 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:52110 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:46547:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52110 dst: /127.0.0.1:46547 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,127 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_480712168_22 at /127.0.0.1:45530 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:37965:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45530 dst: /127.0.0.1:37965 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,127 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:45544 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:37965:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45544 dst: /127.0.0.1:37965 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,128 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:52114 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:46547:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52114 dst: /127.0.0.1:46547 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,128 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_480712168_22 at /127.0.0.1:52076 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:46547:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52076 dst: /127.0.0.1:46547 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,127 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:45550 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:37965:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45550 dst: /127.0.0.1:37965 java.net.SocketException: Connection reset at sun.nio.ch.SocketChannelImpl.throwConnectionReset(SocketChannelImpl.java:394) ~[?:?] at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:426) ~[?:?] at org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,135 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@54926efb{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:00,136 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@31e65952{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:15:00,136 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:15:00,136 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@775e1fb9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:15:00,136 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@54840b6b{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,STOPPED} 2024-12-05T15:15:00,138 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:15:00,138 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:15:00,138 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:15:00,138 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-107838756-172.17.0.2-1733411686024 (Datanode Uuid bde7a88d-0625-402b-9d44-4d4cbb4228a3) service to localhost/127.0.0.1:33521 2024-12-05T15:15:00,139 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data3/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:00,139 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data4/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:00,139 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:15:00,148 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:00,150 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:15:00,151 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:15:00,151 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:15:00,151 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:15:00,152 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@312ce686{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:15:00,152 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@67da9c8c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:15:00,267 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@771dc5f0{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir/jetty-localhost-37999-hadoop-hdfs-3_4_1-tests_jar-_-any-683865880535334559/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:00,268 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4191fe07{HTTP/1.1, (http/1.1)}{localhost:37999} 2024-12-05T15:15:00,268 INFO [Time-limited test {}] server.Server(415): Started @191084ms 2024-12-05T15:15:00,270 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:15:00,283 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:00,287 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1016 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1016 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:00,287 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1015 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1015 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:00,287 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1017 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1017 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:00,288 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:45340 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741834_1010] {}] datanode.DataXceiver(331): 127.0.0.1:37965:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45340 dst: /127.0.0.1:37965 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,288 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:45346 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741833_1009] {}] datanode.DataXceiver(331): 127.0.0.1:37965:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45346 dst: /127.0.0.1:37965 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,289 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_480712168_22 at /127.0.0.1:45352 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741830_1006] {}] datanode.DataXceiver(331): 127.0.0.1:37965:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:45352 dst: /127.0.0.1:37965 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:00,291 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@71404fe1{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:00,291 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@48bfabaa{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:15:00,291 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:15:00,292 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@48245567{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:15:00,292 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@79c9ffa9{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,STOPPED} 2024-12-05T15:15:00,293 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:15:00,293 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-107838756-172.17.0.2-1733411686024 (Datanode Uuid 520a1db3-44dd-4b81-8abe-8a894df71045) service to localhost/127.0.0.1:33521 2024-12-05T15:15:00,293 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:15:00,293 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:15:00,294 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data1/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:00,294 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data2/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:00,295 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:15:00,307 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:00,311 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:15:00,312 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:15:00,312 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:15:00,312 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:15:00,314 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@17c8ba0e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:15:00,314 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4163f715{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:15:00,365 WARN [Thread-1094 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:15:00,368 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x915f9c8d02aff030 with lease ID 0xf7e183916458b2fc: from storage DS-3579be7c-0bfd-45c7-aa32-85fca803b340 node DatanodeRegistration(127.0.0.1:36431, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=41345, infoSecurePort=0, ipcPort=36539, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:00,368 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x915f9c8d02aff030 with lease ID 0xf7e183916458b2fc: from storage DS-736508bc-8897-4ab4-a855-8f92fe91d984 node DatanodeRegistration(127.0.0.1:36431, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=41345, infoSecurePort=0, ipcPort=36539, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:00,434 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@ff463d3{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir/jetty-localhost-44769-hadoop-hdfs-3_4_1-tests_jar-_-any-17918240175277401861/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:00,434 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@4625a0ee{HTTP/1.1, (http/1.1)}{localhost:44769} 2024-12-05T15:15:00,435 INFO [Time-limited test {}] server.Server(415): Started @191250ms 2024-12-05T15:15:00,436 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:15:00,524 WARN [Thread-1125 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:15:00,527 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x208384ffffa7ae65 with lease ID 0xf7e183916458b2fd: from storage DS-c04dc19c-ae7f-4dc6-a523-21295a18894d node DatanodeRegistration(127.0.0.1:43687, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=35285, infoSecurePort=0, ipcPort=41645, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 7, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-05T15:15:00,527 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x208384ffffa7ae65 with lease ID 0xf7e183916458b2fd: from storage DS-1208898a-81c7-4643-a302-3b3a089c1dd8 node DatanodeRegistration(127.0.0.1:43687, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=35285, infoSecurePort=0, ipcPort=41645, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:01,108 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:01,284 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:01,454 INFO [Time-limited test {}] wal.TestLogRolling(349): Data Nodes restarted 2024-12-05T15:15:01,456 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1002 2024-12-05T15:15:01,457 WARN [RS:0;4ecb0f57b706:32857.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=5, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:01,457 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C32857%2C1733411686776:(num 1733411687162) roll requested 2024-12-05T15:15:01,457 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32857 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=5, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:01,458 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:01,458 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32857 {}] ipc.CallRunner(138): callId: 11 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:45020 deadline: 1733411711457, exception=org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=5, requesting roll of WAL 2024-12-05T15:15:01,464 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 newFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:01,464 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=5, requesting roll of WAL 2024-12-05T15:15:01,464 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 with entries=5, filesize=2.09 KB; new WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:01,464 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41345:41345),(127.0.0.1/127.0.0.1:35285:35285)] 2024-12-05T15:15:01,464 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 is not closed yet, will try archiving it next time 2024-12-05T15:15:01,464 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:01,465 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:01,465 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:15:01,465 WARN [IPC Server handler 4 on default port 33521 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 has not been closed. Lease recovery is in progress. RecoveryId = 1019 for block blk_1073741833_1016 2024-12-05T15:15:01,465 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 after 0ms 2024-12-05T15:15:02,108 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:02,284 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:03,109 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:03,285 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:04,110 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:04,286 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:04,368 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741833_1016: GenerationStamp not matched, existing replica is blk_1073741833_1009 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-05T15:15:05,110 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:05,286 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:05,466 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 after 4001ms 2024-12-05T15:15:06,111 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:06,287 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:07,111 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:07,287 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:08,112 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:08,288 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:09,113 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:09,289 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:10,113 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:10,289 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:11,114 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:11,290 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:12,115 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:12,290 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:13,115 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:13,291 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:13,518 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1003 2024-12-05T15:15:14,116 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:14,292 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:15,117 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:15,292 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:15,521 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018 java.io.IOException: Bad response ERROR for BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018 from datanode DatanodeInfoWithStorage[127.0.0.1:43687,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1223) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:15,521 WARN [DataStreamer for file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 block BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018 {}] hdfs.DataStreamer(1731): Error Recovery for BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018 in pipeline [DatanodeInfoWithStorage[127.0.0.1:36431,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK], DatanodeInfoWithStorage[127.0.0.1:43687,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]]: datanode 1(DatanodeInfoWithStorage[127.0.0.1:43687,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]) is bad. 2024-12-05T15:15:15,521 WARN [PacketResponder: BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018, type=HAS_DOWNSTREAM_IN_PIPELINE, downstreams=1:[127.0.0.1:43687] {}] datanode.BlockReceiver$PacketResponder(1529): IOException in PacketResponder.run(): java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.write0(Native Method) ~[?:?] at sun.nio.ch.SocketDispatcher.write(SocketDispatcher.java:62) ~[?:?] at sun.nio.ch.IOUtil.writeFromNativeBuffer(IOUtil.java:132) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:97) ~[?:?] at sun.nio.ch.IOUtil.write(IOUtil.java:53) ~[?:?] at sun.nio.ch.SocketChannelImpl.write(SocketChannelImpl.java:532) ~[?:?] at org.apache.hadoop.net.SocketOutputStream$Writer.performIO(SocketOutputStream.java:62) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:141) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:158) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketOutputStream.write(SocketOutputStream.java:116) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) ~[?:?] at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) ~[?:?] at java.io.DataOutputStream.flush(DataOutputStream.java:128) ~[?:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstreamUnprotected(BlockReceiver.java:1681) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.sendAckUpstream(BlockReceiver.java:1612) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver$PacketResponder.run(BlockReceiver.java:1520) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:15,522 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:50440 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018] {}] datanode.DataXceiver(331): 127.0.0.1:36431:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:50440 dst: /127.0.0.1:36431 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:15,522 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:44496 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018] {}] datanode.DataXceiver(331): 127.0.0.1:43687:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:44496 dst: /127.0.0.1:43687 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:15,523 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@ff463d3{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:15,524 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4625a0ee{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:15:15,524 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:15:15,524 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4163f715{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:15:15,524 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@17c8ba0e{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,STOPPED} 2024-12-05T15:15:15,525 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-107838756-172.17.0.2-1733411686024 (Datanode Uuid 520a1db3-44dd-4b81-8abe-8a894df71045) service to localhost/127.0.0.1:33521 2024-12-05T15:15:15,526 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data1/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:15,526 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data2/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:15,526 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:15:15,544 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:15,548 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:15:15,549 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:15:15,549 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:15:15,549 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:15:15,552 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1164a0b5{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:15:15,553 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@707ec541{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:15:15,667 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@69b93707{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir/jetty-localhost-41485-hadoop-hdfs-3_4_1-tests_jar-_-any-6992112110666136443/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:15,667 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@69edc9c7{HTTP/1.1, (http/1.1)}{localhost:41485} 2024-12-05T15:15:15,667 INFO [Time-limited test {}] server.Server(415): Started @206483ms 2024-12-05T15:15:15,668 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:15:15,686 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1020 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1020 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:15,687 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_-427077916_22 at /127.0.0.1:52682 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741839_1018] {}] datanode.DataXceiver(331): 127.0.0.1:36431:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:52682 dst: /127.0.0.1:36431 java.nio.channels.ClosedChannelException: null at java.nio.channels.spi.AbstractSelectableChannel.register(AbstractSelectableChannel.java:222) ~[?:?] at java.nio.channels.SelectableChannel.register(SelectableChannel.java:260) ~[?:?] at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:334) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:15,691 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@771dc5f0{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:15,692 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@4191fe07{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:15:15,692 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:15:15,692 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@67da9c8c{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:15:15,692 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@312ce686{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,STOPPED} 2024-12-05T15:15:15,694 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:15:15,694 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:15:15,694 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:15:15,694 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-107838756-172.17.0.2-1733411686024 (Datanode Uuid bde7a88d-0625-402b-9d44-4d4cbb4228a3) service to localhost/127.0.0.1:33521 2024-12-05T15:15:15,694 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data3/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:15,695 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data4/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:15,695 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:15:15,703 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:15,707 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:15:15,712 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:15:15,712 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:15:15,712 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:15:15,713 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@390622b4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:15:15,713 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1b9b7f77{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:15:15,757 WARN [Thread-1169 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:15:15,760 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x74b828efe93bd59b with lease ID 0xf7e183916458b2fe: from storage DS-c04dc19c-ae7f-4dc6-a523-21295a18894d node DatanodeRegistration(127.0.0.1:45947, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=33007, infoSecurePort=0, ipcPort=40283, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:15,760 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x74b828efe93bd59b with lease ID 0xf7e183916458b2fe: from storage DS-1208898a-81c7-4643-a302-3b3a089c1dd8 node DatanodeRegistration(127.0.0.1:45947, datanodeUuid=520a1db3-44dd-4b81-8abe-8a894df71045, infoPort=33007, infoSecurePort=0, ipcPort=40283, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 8, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:15,831 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@30ab6650{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/java.io.tmpdir/jetty-localhost-39599-hadoop-hdfs-3_4_1-tests_jar-_-any-14703207837772099655/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:15,831 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@508acd9e{HTTP/1.1, (http/1.1)}{localhost:39599} 2024-12-05T15:15:15,831 INFO [Time-limited test {}] server.Server(415): Started @206647ms 2024-12-05T15:15:15,832 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:15:15,928 WARN [Thread-1200 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:15:15,930 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbb51a42c29a3ad83 with lease ID 0xf7e183916458b2ff: from storage DS-3579be7c-0bfd-45c7-aa32-85fca803b340 node DatanodeRegistration(127.0.0.1:46879, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=39823, infoSecurePort=0, ipcPort=33481, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 8, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:15,931 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xbb51a42c29a3ad83 with lease ID 0xf7e183916458b2ff: from storage DS-736508bc-8897-4ab4-a855-8f92fe91d984 node DatanodeRegistration(127.0.0.1:46879, datanodeUuid=bde7a88d-0625-402b-9d44-4d4cbb4228a3, infoPort=39823, infoSecurePort=0, ipcPort=33481, storageInfo=lv=-57;cid=testClusterID;nsid=884883054;c=1733411686024), blocks: 7, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:16,117 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:16,293 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:16,704 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:15:16,853 INFO [Time-limited test {}] wal.TestLogRolling(366): Data Nodes restarted 2024-12-05T15:15:16,854 INFO [Time-limited test {}] wal.AbstractTestLogRolling(330): Validated row row1004 2024-12-05T15:15:16,855 WARN [RS:0;4ecb0f57b706:32857.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=8, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36431,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,855 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C32857%2C1733411686776:(num 1733411701457) roll requested 2024-12-05T15:15:16,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32857 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=8, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36431,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,856 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C32857%2C1733411686776.1733411716855 2024-12-05T15:15:16,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=32857 {}] ipc.CallRunner(138): callId: 18 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:45020 deadline: 1733411726855, exception=org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=8, requesting roll of WAL 2024-12-05T15:15:16,862 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 newFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 2024-12-05T15:15:16,862 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=8, requesting roll of WAL 2024-12-05T15:15:16,862 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 with entries=2, filesize=2.37 KB; new WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 2024-12-05T15:15:16,862 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39823:39823),(127.0.0.1/127.0.0.1:33007:33007)] 2024-12-05T15:15:16,862 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 is not closed yet, will try archiving it next time 2024-12-05T15:15:16,862 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36431,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,862 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:36431,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,862 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:16,863 WARN [IPC Server handler 3 on default port 33521 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 has not been closed. Lease recovery is in progress. RecoveryId = 1022 for block blk_1073741839_1020 2024-12-05T15:15:16,863 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 after 1ms 2024-12-05T15:15:16,923 WARN [master/4ecb0f57b706:0:becomeActiveMaster.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=95, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,923 DEBUG [master:store-WAL-Roller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C34185%2C1733411686721:(num 1733411686854) roll requested 2024-12-05T15:15:16,924 ERROR [ProcExecTimeout {}] region.RegionProcedureStore(422): Failed to delete pids=[4, 7, 8, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,924 INFO [master:store-WAL-Roller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C34185%2C1733411686721.1733411716924 2024-12-05T15:15:16,924 ERROR [ProcExecTimeout {}] procedure2.TimeoutExecutorThread(124): Ignoring pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner exception: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL java.io.UncheckedIOException: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL at org.apache.hadoop.hbase.procedure2.store.region.RegionProcedureStore.delete(RegionProcedureStore.java:423) ~[classes/:?] at org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner.periodicExecute(CompletedProcedureCleaner.java:135) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.executeInMemoryChore(TimeoutExecutorThread.java:122) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.execDelayedProcedure(TimeoutExecutorThread.java:101) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.procedure2.TimeoutExecutorThread.run(TimeoutExecutorThread.java:68) ~[hbase-procedure-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] Caused by: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,929 WARN [master:store-WAL-Roller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=95, requesting roll of WAL 2024-12-05T15:15:16,930 INFO [master:store-WAL-Roller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 with entries=92, filesize=46.00 KB; new WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411716924 2024-12-05T15:15:16,930 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39823:39823),(127.0.0.1/127.0.0.1:33007:33007)] 2024-12-05T15:15:16,930 DEBUG [master:store-WAL-Roller {}] wal.AbstractFSWAL(751): hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 is not closed yet, will try archiving it next time 2024-12-05T15:15:16,930 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,930 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:16,930 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 2024-12-05T15:15:16,931 WARN [IPC Server handler 4 on default port 33521 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 has not been closed. Lease recovery is in progress. RecoveryId = 1024 for block blk_1073741830_1017 2024-12-05T15:15:16,931 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 after 1ms 2024-12-05T15:15:17,118 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:17,293 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:18,119 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:18,294 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:18,882 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 817a20dfbf0d1531445f95325bcaf27c, had cached 0 bytes from a total of 23930 2024-12-05T15:15:19,119 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:19,295 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:19,931 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741830_1017: GenerationStamp not matched, existing replica is blk_1073741830_1006 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-05T15:15:20,120 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:20,295 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:20,864 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 after 4002ms 2024-12-05T15:15:20,931 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721/4ecb0f57b706%2C34185%2C1733411686721.1733411686854 after 4001ms 2024-12-05T15:15:21,010 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:15:21,011 INFO [RS-EventLoopGroup-8-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:38066, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.3 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:15:21,121 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:21,296 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:22,121 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:22,297 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:22,759 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741839_1020: GenerationStamp not matched, existing replica is blk_1073741839_1018 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-05T15:15:23,122 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:23,297 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:24,122 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:24,298 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:25,123 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:25,124 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=2 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 after 68044ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:25,298 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:26,124 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:26,299 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:27,125 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:27,300 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:28,125 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:28,300 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:28,458 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-05T15:15:28,458 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-05T15:15:28,901 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:28,907 DEBUG [Time-limited test {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 newFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:28,908 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 with entries=1, filesize=1.23 KB; new WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:28,908 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39823:39823),(127.0.0.1/127.0.0.1:33007:33007)] 2024-12-05T15:15:28,908 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 is not closed yet, will try archiving it next time 2024-12-05T15:15:28,909 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:15:28,909 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:15:28,909 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 after 0ms 2024-12-05T15:15:28,909 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:15:28,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741840_1021 (size=1264) 2024-12-05T15:15:28,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741840_1021 (size=1264) 2024-12-05T15:15:28,916 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #3: [\x00/METAFAMILY:HBASE::REGION_EVENT::REGION_OPEN/1733411687926/Put/vlen=162/seqid=0] 2024-12-05T15:15:28,916 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #4: [default/info:d/1733411687988/Put/vlen=9/seqid=0] 2024-12-05T15:15:28,916 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #5: [hbase/info:d/1733411688009/Put/vlen=7/seqid=0] 2024-12-05T15:15:28,916 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #3: [\x00/METAFAMILY:HBASE::REGION_EVENT::REGION_OPEN/1733411688464/Put/vlen=218/seqid=0] 2024-12-05T15:15:28,916 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #4: [row1002/info:/1733411698123/Put/vlen=1045/seqid=0] 2024-12-05T15:15:28,917 DEBUG [Time-limited test {}] wal.TestLogRolling(396): EOF reading file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411687162 2024-12-05T15:15:28,917 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:28,917 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:28,917 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 after 0ms 2024-12-05T15:15:28,917 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:28,920 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #6: [row1003/info:/1733411711515/Put/vlen=1045/seqid=0] 2024-12-05T15:15:28,920 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #7: [row1004/info:/1733411713519/Put/vlen=1045/seqid=0] 2024-12-05T15:15:28,920 DEBUG [Time-limited test {}] wal.TestLogRolling(396): EOF reading file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411701457 2024-12-05T15:15:28,920 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 2024-12-05T15:15:28,920 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 2024-12-05T15:15:28,921 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 after 1ms 2024-12-05T15:15:28,921 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411716855 2024-12-05T15:15:28,924 DEBUG [Time-limited test {}] wal.TestLogRolling(389): #9: [row1005/info:/1733411726900/Put/vlen=1045/seqid=0] 2024-12-05T15:15:28,924 DEBUG [Time-limited test {}] wal.TestLogRolling(380): recovering lease for hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:28,924 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:28,924 WARN [IPC Server handler 4 on default port 33521 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 has not been closed. Lease recovery is in progress. RecoveryId = 1026 for block blk_1073741842_1025 2024-12-05T15:15:28,924 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 after 0ms 2024-12-05T15:15:29,126 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:29,301 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:29,763 WARN [ResponseProcessor for block BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 {}] hdfs.DataStreamer$ResponseProcessor(1303): Exception for BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 java.io.EOFException: Unexpected EOF while trying to read response from server at org.apache.hadoop.hdfs.protocolPB.PBHelperClient.vintPrefixed(PBHelperClient.java:529) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck.readFields(PipelineAck.java:244) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer$ResponseProcessor.run(DataStreamer.java:1180) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:29,763 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_480712168_22 at /127.0.0.1:35148 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025] {}] datanode.DataXceiver(331): 127.0.0.1:46879:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:35148 dst: /127.0.0.1:46879 java.io.InterruptedIOException: Interrupted while waiting for IO on channel java.nio.channels.SocketChannel[connected local=localhost/127.0.0.1:46879 remote=/127.0.0.1:35148]. Total timeout mills is 60000, 59144 millis timeout left. at org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:350) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) ~[hadoop-common-3.4.1.jar:?] at java.io.BufferedInputStream.fill(BufferedInputStream.java:244) ~[?:?] at java.io.BufferedInputStream.read1(BufferedInputStream.java:284) ~[?:?] at java.io.BufferedInputStream.read(BufferedInputStream.java:343) ~[?:?] at java.io.DataInputStream.read(DataInputStream.java:151) ~[?:?] at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:214) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:29,763 ERROR [DataXceiver for client DFSClient_NONMAPREDUCE_480712168_22 at /127.0.0.1:39074 [Receiving block BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025] {}] datanode.DataXceiver(331): 127.0.0.1:45947:DataXceiver error processing WRITE_BLOCK operation src: /127.0.0.1:39074 dst: /127.0.0.1:45947 java.io.IOException: Premature EOF from inputStream at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:216) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:221) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:144) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:119) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:553) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:1011) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:920) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:176) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:110) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:299) ~[hadoop-hdfs-3.4.1.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:29,763 WARN [DataStreamer for file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 block BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 {}] hdfs.DataStreamer(1731): Error Recovery for BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 in pipeline [DatanodeInfoWithStorage[127.0.0.1:46879,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK], DatanodeInfoWithStorage[127.0.0.1:45947,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]]: datanode 0(DatanodeInfoWithStorage[127.0.0.1:46879,DS-3579be7c-0bfd-45c7-aa32-85fca803b340,DISK]) is bad. 2024-12-05T15:15:29,767 WARN [DataStreamer for file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 block BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 {}] hdfs.DataStreamer(859): DataStreamer Exception org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:29,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741842_1026 (size=85) 2024-12-05T15:15:30,127 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:30,301 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:31,127 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:31,302 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:32,128 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:32,303 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:32,925 INFO [Time-limited test {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 after 4001ms 2024-12-05T15:15:32,925 DEBUG [Time-limited test {}] wal.TestLogRolling(384): Reading WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:32,929 DEBUG [Time-limited test {}] wal.TestLogRolling(396): EOF reading file /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:32,930 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 63c9f1b847ea3792bcfec0f2259fa595 1/1 column families, dataSize=4.20 KB heapSize=4.75 KB 2024-12-05T15:15:32,930 WARN [RS:0;4ecb0f57b706:32857.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=11, requesting roll of WAL org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,930 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C32857%2C1733411686776:(num 1733411728901) roll requested 2024-12-05T15:15:32,930 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 63c9f1b847ea3792bcfec0f2259fa595: 2024-12-05T15:15:32,931 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C32857%2C1733411686776.1733411732930 2024-12-05T15:15:32,931 INFO [Time-limited test {}] wal.TestLogRolling(416): org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=11, requesting roll of WAL org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=11, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,931 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 9a5acd1c15e040330ff7ee9cb3eb52c2 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-05T15:15:32,931 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 9a5acd1c15e040330ff7ee9cb3eb52c2: 2024-12-05T15:15:32,931 INFO [Time-limited test {}] wal.TestLogRolling(416): org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=11, requesting roll of WAL org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=11, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,932 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.90 KB heapSize=5.42 KB 2024-12-05T15:15:32,932 WARN [RS_OPEN_META-regionserver/4ecb0f57b706:0-0.append-pool-0 {}] wal.FSHLog$RingBufferEventHandler(1189): Append sequenceId=15, requesting roll of WAL java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,932 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 1588230740: 2024-12-05T15:15:32,933 INFO [Time-limited test {}] wal.TestLogRolling(416): org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,935 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-05T15:15:32,935 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-05T15:15:32,935 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1ed3dd55 to 127.0.0.1:50082 2024-12-05T15:15:32,935 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:15:32,935 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T15:15:32,935 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=1723237142, stopped=false 2024-12-05T15:15:32,935 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=4ecb0f57b706,34185,1733411686721 2024-12-05T15:15:32,937 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.TestLogRolling$2(324): preLogRoll: oldFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 newFile=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411732930 2024-12-05T15:15:32,937 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=11, requesting roll of WAL 2024-12-05T15:15:32,937 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411732930 2024-12-05T15:15:32,937 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:33007:33007),(127.0.0.1/127.0.0.1:39823:39823)] 2024-12-05T15:15:32,938 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 is not closed yet, will try archiving it next time 2024-12-05T15:15:32,938 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(197): WAL FSHLog 4ecb0f57b706%2C32857%2C1733411686776.meta:.meta(num 1733411687525) roll requested 2024-12-05T15:15:32,937 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,938 INFO [regionserver/4ecb0f57b706:0.logRoller {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C32857%2C1733411686776.meta.1733411732938.meta 2024-12-05T15:15:32,938 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. org.apache.hadoop.ipc.RemoteException: Unexpected BlockUCState: BP-107838756-172.17.0.2-1733411686024:blk_1073741842_1025 is UNDER_RECOVERY but not UNDER_CONSTRUCTION at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkUCBlock(FSNamesystem.java:5912) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.bumpBlockGenerationStamp(FSNamesystem.java:5980) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.updateBlockForPipeline(NameNodeRpcServer.java:1002) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolServerSideTranslatorPB.java:1182) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3198) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1529) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.Client.call(Client.java:1426) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy45.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$updateBlockForPipeline$50(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.updateBlockForPipeline(ClientNamenodeProtocolTranslatorPB.java:920) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) ~[hadoop-common-3.4.1.jar:?] at jdk.proxy2.$Proxy46.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at jdk.internal.reflect.GeneratedMethodAccessor114.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.fs.HFileSystem$1.invoke(HFileSystem.java:363) ~[classes/:?] at jdk.proxy2.$Proxy47.updateBlockForPipeline(Unknown Source) ~[?:?] at org.apache.hadoop.hdfs.DataStreamer.updateBlockForPipeline(DataStreamer.java:1786) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1657) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,938 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:32,938 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:15:32,938 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-05T15:15:32,938 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:32,939 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:15:32,939 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 after 1ms 2024-12-05T15:15:32,939 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,32857,1733411686776' ***** 2024-12-05T15:15:32,939 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-05T15:15:32,939 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T15:15:32,939 INFO [RS:0;4ecb0f57b706:32857 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T15:15:32,939 INFO [RS:0;4ecb0f57b706:32857 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T15:15:32,939 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:15:32,939 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(3579): Received CLOSE for 63c9f1b847ea3792bcfec0f2259fa595 2024-12-05T15:15:32,939 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-05T15:15:32,940 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:15:32,940 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:32,940 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:15:32,941 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 to hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/oldWALs/4ecb0f57b706%2C32857%2C1733411686776.1733411728901 2024-12-05T15:15:32,941 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(3579): Received CLOSE for 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:15:32,941 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,32857,1733411686776 2024-12-05T15:15:32,941 DEBUG [RS:0;4ecb0f57b706:32857 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:15:32,941 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 63c9f1b847ea3792bcfec0f2259fa595, disabling compactions & flushes 2024-12-05T15:15:32,941 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T15:15:32,941 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:15:32,941 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T15:15:32,941 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T15:15:32,941 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:15:32,941 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. after waiting 0 ms 2024-12-05T15:15:32,941 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:15:32,941 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:15:32,941 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 63c9f1b847ea3792bcfec0f2259fa595 1/1 column families, dataSize=4.20 KB heapSize=4.98 KB 2024-12-05T15:15:32,941 WARN [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultMemStore(92): Snapshot called again without clearing previous. Doing nothing. Another ongoing flush or did we fail last attempt? 2024-12-05T15:15:32,942 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-05T15:15:32,942 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1603): Online Regions={63c9f1b847ea3792bcfec0f2259fa595=TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595., 9a5acd1c15e040330ff7ee9cb3eb52c2=hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2., 1588230740=hbase:meta,,1.1588230740} 2024-12-05T15:15:32,942 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:15:32,942 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:15:32,942 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 63c9f1b847ea3792bcfec0f2259fa595, 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:15:32,942 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:15:32,942 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:15:32,942 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:15:32,942 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.90 KB heapSize=5.89 KB 2024-12-05T15:15:32,942 WARN [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultMemStore(92): Snapshot called again without clearing previous. Doing nothing. Another ongoing flush or did we fail last attempt? 2024-12-05T15:15:32,942 WARN [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultMemStore(92): Snapshot called again without clearing previous. Doing nothing. Another ongoing flush or did we fail last attempt? 2024-12-05T15:15:32,943 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:15:32,943 ERROR [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2808): ***** ABORTING region server 4ecb0f57b706,32857,1733411686776: Unrecoverable exception while closing hbase:meta,,1.1588230740 ***** org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) ~[classes/:?] at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) ~[disruptor-3.4.4.jar:?] at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) ~[disruptor-3.4.4.jar:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,943 ERROR [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2815): RegionServer abort: loaded coprocessors are: [org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint] 2024-12-05T15:15:32,943 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for java.lang:type=Memory 2024-12-05T15:15:32,944 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=IPC 2024-12-05T15:15:32,944 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=Replication 2024-12-05T15:15:32,944 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] util.JSONBean(135): Listing beans for Hadoop:service=HBase,name=RegionServer,sub=Server 2024-12-05T15:15:32,944 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2819): Dump of metrics as JSON on abort: { "beans": [ { "name": "java.lang:type=Memory", "modelerType": "sun.management.MemoryImpl", "ObjectPendingFinalizationCount": 0, "HeapMemoryUsage": { "committed": 1048576000, "init": 1048576000, "max": 2306867200, "used": 669697520 }, "NonHeapMemoryUsage": { "committed": 169541632, "init": 7667712, "max": -1, "used": 167619304 }, "Verbose": false, "ObjectName": "java.lang:type=Memory" } ], "beans": [], "beans": [], "beans": [] } 2024-12-05T15:15:32,945 WARN [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=34185 {}] master.MasterRpcServices(626): 4ecb0f57b706,32857,1733411686776 reported a fatal error: ***** ABORTING region server 4ecb0f57b706,32857,1733411686776: Unrecoverable exception while closing hbase:meta,,1.1588230740 ***** Cause: org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1191) at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1064) at org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:967) at com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:168) at com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) at java.base/java.lang.Thread.run(Thread.java:840) Caused by: java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) 2024-12-05T15:15:32,953 WARN [regionserver/4ecb0f57b706:0.logRoller {}] wal.FSHLog(373): Failed sync-before-close but no outstanding appends; closing WALorg.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append sequenceId=15, requesting roll of WAL 2024-12-05T15:15:32,953 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta with entries=11, filesize=3.66 KB; new WAL /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411732938.meta 2024-12-05T15:15:32,955 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39823:39823),(127.0.0.1/127.0.0.1:33007:33007)] 2024-12-05T15:15:32,955 DEBUG [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractFSWAL(751): hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta is not closed yet, will try archiving it next time 2024-12-05T15:15:32,955 WARN [Close-WAL-Writer-0 {}] wal.AbstractProtobufLogWriter(255): Failed to write trailer, non-fatal, continuing... java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,955 WARN [Close-WAL-Writer-0 {}] wal.FSHLog(462): close old writer failed. java.io.IOException: All datanodes [DatanodeInfoWithStorage[127.0.0.1:37965,DS-c04dc19c-ae7f-4dc6-a523-21295a18894d,DISK]] are bad. Aborting... at org.apache.hadoop.hdfs.DataStreamer.handleBadDatanode(DataStreamer.java:1721) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1644) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1627) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.processDatanodeOrExternalError(DataStreamer.java:1408) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:707) ~[hadoop-hdfs-client-3.4.1.jar:?] 2024-12-05T15:15:32,956 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta 2024-12-05T15:15:32,956 WARN [IPC Server handler 3 on default port 33521 {}] namenode.FSNamesystem(3854): DIR* NameSystem.internalReleaseLease: File /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta has not been closed. Lease recovery is in progress. RecoveryId = 1029 for block blk_1073741834_1015 2024-12-05T15:15:32,956 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Failed to recover lease, attempt=0 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta after 0ms 2024-12-05T15:15:32,965 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595/.tmp/info/5f483053999b46d9b89695d91e53be8e is 1080, key is row1002/info:/1733411698123/Put/seqid=0 2024-12-05T15:15:32,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741845_1030 (size=9270) 2024-12-05T15:15:32,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741845_1030 (size=9270) 2024-12-05T15:15:32,971 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=4.20 KB at sequenceid=12 (bloomFilter=true), to=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595/.tmp/info/5f483053999b46d9b89695d91e53be8e 2024-12-05T15:15:32,978 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595/.tmp/info/5f483053999b46d9b89695d91e53be8e as hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595/info/5f483053999b46d9b89695d91e53be8e 2024-12-05T15:15:32,983 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595/info/5f483053999b46d9b89695d91e53be8e, entries=4, sequenceid=12, filesize=9.1 K 2024-12-05T15:15:32,984 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~4.20 KB/4304, heapSize ~4.73 KB/4848, currentSize=0 B/0 for 63c9f1b847ea3792bcfec0f2259fa595 in 43ms, sequenceid=12, compaction requested=false 2024-12-05T15:15:32,988 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/data/default/TestLogRolling-testLogRollOnPipelineRestart/63c9f1b847ea3792bcfec0f2259fa595/recovered.edits/15.seqid, newMaxSeqId=15, maxSeqId=1 2024-12-05T15:15:32,989 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:15:32,989 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 63c9f1b847ea3792bcfec0f2259fa595: 2024-12-05T15:15:32,989 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRollOnPipelineRestart,,1733411688110.63c9f1b847ea3792bcfec0f2259fa595. 2024-12-05T15:15:32,989 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 9a5acd1c15e040330ff7ee9cb3eb52c2, disabling compactions & flushes 2024-12-05T15:15:32,989 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:32,989 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:32,989 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. after waiting 0 ms 2024-12-05T15:15:32,989 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:32,990 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 9a5acd1c15e040330ff7ee9cb3eb52c2: 2024-12-05T15:15:32,990 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:33,025 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-05T15:15:33,025 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-05T15:15:33,029 INFO [regionserver/4ecb0f57b706:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:15:33,128 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:33,142 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(3579): Received CLOSE for 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:15:33,142 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:15:33,142 DEBUG [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 9a5acd1c15e040330ff7ee9cb3eb52c2 2024-12-05T15:15:33,142 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 9a5acd1c15e040330ff7ee9cb3eb52c2, disabling compactions & flushes 2024-12-05T15:15:33,142 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:33,142 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:15:33,142 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. after waiting 0 ms 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:33,143 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 9a5acd1c15e040330ff7ee9cb3eb52c2: 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing hbase:namespace,,1733411687573.9a5acd1c15e040330ff7ee9cb3eb52c2. 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:15:33,143 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionServer(2803): Abort already in progress. Ignoring the current request with reason: Unrecoverable exception while closing hbase:meta,,1.1588230740 2024-12-05T15:15:33,303 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:33,343 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1624): We were exiting though online regions are not empty, because some regions failed closing 2024-12-05T15:15:33,343 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,32857,1733411686776; all regions closed. 2024-12-05T15:15:33,343 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776 2024-12-05T15:15:33,345 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741844_1028 (size=93) 2024-12-05T15:15:33,345 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741844_1028 (size=93) 2024-12-05T15:15:34,129 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:34,304 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:34,930 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1452): Error processing datanode Command java.io.IOException: Failed to delete 1 (out of 1) replica(s): 0) Failed to delete replica blk_1073741834_1015: GenerationStamp not matched, existing replica is blk_1073741834_1010 at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2389) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.FsDatasetImpl.invalidate(FsDatasetImpl.java:2317) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActive(BPOfferService.java:743) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPOfferService.processCommandFromActor(BPOfferService.java:692) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processCommand(BPServiceActor.java:1439) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.lambda$enqueue$2(BPServiceActor.java:1485) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1412) ~[hadoop-hdfs-3.4.1.jar:?] at org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) ~[hadoop-hdfs-3.4.1.jar:?] 2024-12-05T15:15:34,991 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRollOnPipelineRestart 2024-12-05T15:15:35,130 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:35,304 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:36,130 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:36,305 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:36,957 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(223): Recovered lease, attempt=1 on file=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776/4ecb0f57b706%2C32857%2C1733411686776.meta.1733411687525.meta after 4001ms 2024-12-05T15:15:36,958 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/WALs/4ecb0f57b706,32857,1733411686776 2024-12-05T15:15:36,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741843_1027 (size=1162) 2024-12-05T15:15:36,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741843_1027 (size=1162) 2024-12-05T15:15:36,961 DEBUG [RS:0;4ecb0f57b706:32857 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:15:36,961 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:15:36,961 INFO [RS:0;4ecb0f57b706:32857 {}] hbase.ChoreService(370): Chore service for: regionserver/4ecb0f57b706:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-05T15:15:36,961 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:15:36,961 INFO [RS:0;4ecb0f57b706:32857 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:32857 2024-12-05T15:15:36,963 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/4ecb0f57b706,32857,1733411686776 2024-12-05T15:15:36,963 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:15:36,965 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [4ecb0f57b706,32857,1733411686776] 2024-12-05T15:15:36,965 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 4ecb0f57b706,32857,1733411686776; numProcessing=1 2024-12-05T15:15:36,966 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/4ecb0f57b706,32857,1733411686776 already deleted, retry=false 2024-12-05T15:15:36,966 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 4ecb0f57b706,32857,1733411686776 expired; onlineServers=0 2024-12-05T15:15:36,966 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,34185,1733411686721' ***** 2024-12-05T15:15:36,966 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T15:15:36,966 DEBUG [M:0;4ecb0f57b706:34185 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@58453820, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:15:36,966 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,34185,1733411686721 2024-12-05T15:15:36,966 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,34185,1733411686721; all regions closed. 2024-12-05T15:15:36,966 DEBUG [M:0;4ecb0f57b706:34185 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:15:36,966 DEBUG [M:0;4ecb0f57b706:34185 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T15:15:36,966 DEBUG [M:0;4ecb0f57b706:34185 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T15:15:36,966 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T15:15:36,966 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411686924 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411686924,5,FailOnTimeoutGroup] 2024-12-05T15:15:36,966 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411686924 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411686924,5,FailOnTimeoutGroup] 2024-12-05T15:15:36,967 INFO [M:0;4ecb0f57b706:34185 {}] hbase.ChoreService(370): Chore service for: master/4ecb0f57b706:0 had [] on shutdown 2024-12-05T15:15:36,967 DEBUG [M:0;4ecb0f57b706:34185 {}] master.HMaster(1733): Stopping service threads 2024-12-05T15:15:36,967 INFO [M:0;4ecb0f57b706:34185 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T15:15:36,967 INFO [M:0;4ecb0f57b706:34185 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T15:15:36,967 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T15:15:36,968 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T15:15:36,968 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:36,968 DEBUG [M:0;4ecb0f57b706:34185 {}] zookeeper.ZKUtil(347): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T15:15:36,968 WARN [M:0;4ecb0f57b706:34185 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T15:15:36,968 INFO [M:0;4ecb0f57b706:34185 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-05T15:15:36,968 INFO [M:0;4ecb0f57b706:34185 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T15:15:36,968 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:15:36,968 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:15:36,968 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:36,968 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:36,968 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:15:36,968 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:36,969 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=40.10 KB heapSize=49.26 KB 2024-12-05T15:15:36,991 DEBUG [M:0;4ecb0f57b706:34185 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/80568265447743e8827e2533c2c80dad is 82, key is hbase:meta,,1/info:regioninfo/1733411687551/Put/seqid=0 2024-12-05T15:15:36,996 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741846_1031 (size=5672) 2024-12-05T15:15:36,996 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741846_1031 (size=5672) 2024-12-05T15:15:36,996 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/80568265447743e8827e2533c2c80dad 2024-12-05T15:15:37,018 DEBUG [M:0;4ecb0f57b706:34185 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/bb6edecb044a4ec7b7f3ec8627b246d9 is 779, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733411688481/Put/seqid=0 2024-12-05T15:15:37,023 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741847_1032 (size=7470) 2024-12-05T15:15:37,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741847_1032 (size=7470) 2024-12-05T15:15:37,024 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=39.50 KB at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/bb6edecb044a4ec7b7f3ec8627b246d9 2024-12-05T15:15:37,044 DEBUG [M:0;4ecb0f57b706:34185 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/b91e6b8b55164d11bdd0e34dba770fc4 is 69, key is 4ecb0f57b706,32857,1733411686776/rs:state/1733411687013/Put/seqid=0 2024-12-05T15:15:37,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741848_1033 (size=5156) 2024-12-05T15:15:37,049 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741848_1033 (size=5156) 2024-12-05T15:15:37,050 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/b91e6b8b55164d11bdd0e34dba770fc4 2024-12-05T15:15:37,065 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:15:37,065 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:32857-0x10062bf02190001, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:15:37,065 INFO [RS:0;4ecb0f57b706:32857 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,32857,1733411686776; zookeeper connection closed. 2024-12-05T15:15:37,065 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@1bf407c1 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@1bf407c1 2024-12-05T15:15:37,066 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-05T15:15:37,069 DEBUG [M:0;4ecb0f57b706:34185 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/a308b48fbd454d41b6ec7e1d6526fd77 is 52, key is load_balancer_on/state:d/1733411688104/Put/seqid=0 2024-12-05T15:15:37,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741849_1034 (size=5056) 2024-12-05T15:15:37,076 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741849_1034 (size=5056) 2024-12-05T15:15:37,076 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=96 (bloomFilter=true), to=hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/a308b48fbd454d41b6ec7e1d6526fd77 2024-12-05T15:15:37,081 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/80568265447743e8827e2533c2c80dad as hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/80568265447743e8827e2533c2c80dad 2024-12-05T15:15:37,087 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/80568265447743e8827e2533c2c80dad, entries=8, sequenceid=96, filesize=5.5 K 2024-12-05T15:15:37,088 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/bb6edecb044a4ec7b7f3ec8627b246d9 as hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/bb6edecb044a4ec7b7f3ec8627b246d9 2024-12-05T15:15:37,093 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/bb6edecb044a4ec7b7f3ec8627b246d9, entries=11, sequenceid=96, filesize=7.3 K 2024-12-05T15:15:37,094 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/b91e6b8b55164d11bdd0e34dba770fc4 as hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/b91e6b8b55164d11bdd0e34dba770fc4 2024-12-05T15:15:37,099 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/b91e6b8b55164d11bdd0e34dba770fc4, entries=1, sequenceid=96, filesize=5.0 K 2024-12-05T15:15:37,100 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/a308b48fbd454d41b6ec7e1d6526fd77 as hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/a308b48fbd454d41b6ec7e1d6526fd77 2024-12-05T15:15:37,104 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:33521/user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/a308b48fbd454d41b6ec7e1d6526fd77, entries=1, sequenceid=96, filesize=4.9 K 2024-12-05T15:15:37,105 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(3040): Finished flush of dataSize ~40.10 KB/41064, heapSize ~49.20 KB/50376, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 137ms, sequenceid=96, compaction requested=false 2024-12-05T15:15:37,107 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:37,107 DEBUG [M:0;4ecb0f57b706:34185 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:15:37,107 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/289645fc-61d7-d25d-c9ed-dab31dd59a0b/MasterData/WALs/4ecb0f57b706,34185,1733411686721 2024-12-05T15:15:37,109 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45947 is added to blk_1073741841_1023 (size=757) 2024-12-05T15:15:37,109 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46879 is added to blk_1073741841_1023 (size=757) 2024-12-05T15:15:37,110 INFO [M:0;4ecb0f57b706:34185 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-05T15:15:37,110 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:15:37,110 INFO [M:0;4ecb0f57b706:34185 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:34185 2024-12-05T15:15:37,112 DEBUG [M:0;4ecb0f57b706:34185 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/4ecb0f57b706,34185,1733411686721 already deleted, retry=false 2024-12-05T15:15:37,131 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:37,214 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:15:37,214 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:34185-0x10062bf02190000, quorum=127.0.0.1:50082, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:15:37,214 INFO [M:0;4ecb0f57b706:34185 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,34185,1733411686721; zookeeper connection closed. 2024-12-05T15:15:37,218 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@30ab6650{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:37,219 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@508acd9e{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:15:37,219 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:15:37,219 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1b9b7f77{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:15:37,219 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@390622b4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,STOPPED} 2024-12-05T15:15:37,221 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:15:37,221 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:15:37,221 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:15:37,221 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-107838756-172.17.0.2-1733411686024 (Datanode Uuid bde7a88d-0625-402b-9d44-4d4cbb4228a3) service to localhost/127.0.0.1:33521 2024-12-05T15:15:37,222 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data3/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:37,222 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data4/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:37,222 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:15:37,224 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@69b93707{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:37,225 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@69edc9c7{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:15:37,225 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:15:37,225 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@707ec541{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:15:37,225 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1164a0b5{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,STOPPED} 2024-12-05T15:15:37,227 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:15:37,227 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:15:37,228 WARN [BP-107838756-172.17.0.2-1733411686024 heartbeating to localhost/127.0.0.1:33521 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-107838756-172.17.0.2-1733411686024 (Datanode Uuid 520a1db3-44dd-4b81-8abe-8a894df71045) service to localhost/127.0.0.1:33521 2024-12-05T15:15:37,228 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:15:37,228 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data1/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:37,229 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/cluster_9b60ad6b-f2bd-0209-bed3-62bc6688c09c/dfs/data/data2/current/BP-107838756-172.17.0.2-1733411686024 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:15:37,229 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:15:37,238 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@29c97027{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:15:37,238 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@38cff363{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:15:37,238 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:15:37,238 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@134868e9{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:15:37,239 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@64c90eeb{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir/,STOPPED} 2024-12-05T15:15:37,246 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-05T15:15:37,265 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-05T15:15:37,273 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnPipelineRestart Thread=101 (was 86) Potentially hanging thread: RS-EventLoopGroup-8-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins@localhost:33521 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-9-2 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33521 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LeaseRenewer:jenkins.hfs.3@localhost:33521 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-8-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-26-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-28-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-26-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-27-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-29-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-6 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-8-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-27-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-26-2 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-29-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Command processor java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Potentially hanging thread: nioEventLoopGroup-29-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:33521 from jenkins.hfs.3 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-28-1 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:33521 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: nioEventLoopGroup-28-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33521 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-9-1 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:33521 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-4-5 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-9-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: nioEventLoopGroup-27-3 java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//io.netty.util.concurrent.SingleThreadEventExecutor.confirmShutdown(SingleThreadEventExecutor.java:787) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:596) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (1454462474) connection to localhost/127.0.0.1:33521 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) - Thread LEAK? -, OpenFileDescriptor=444 (was 414) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=130 (was 168), ProcessCount=11 (was 11), AvailableMemoryMB=8568 (was 8674) 2024-12-05T15:15:37,280 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testCompactionRecordDoesntBlockRolling Thread=101, OpenFileDescriptor=444, MaxFileDescriptor=1048576, SystemLoadAverage=130, ProcessCount=11, AvailableMemoryMB=8568 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.log.dir so I do NOT create it in target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/5740d6fd-6ada-c5fd-4833-159420a2dc18/hadoop.tmp.dir so I do NOT create it in target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6, deleteOnExit=true 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/test.cache.data in system properties and HBase conf 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir in system properties and HBase conf 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T15:15:37,281 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-05T15:15:37,281 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/nfs.dump.dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/java.io.tmpdir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:15:37,282 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T15:15:37,283 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T15:15:37,296 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:15:37,306 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:37,370 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:37,375 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:15:37,377 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:15:37,377 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:15:37,377 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:15:37,377 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:37,378 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7cea72ca{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:15:37,378 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5559bf37{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:15:37,492 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@307f11b9{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/java.io.tmpdir/jetty-localhost-35579-hadoop-hdfs-3_4_1-tests_jar-_-any-8474723366419887457/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:15:37,493 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6bcef121{HTTP/1.1, (http/1.1)}{localhost:35579} 2024-12-05T15:15:37,493 INFO [Time-limited test {}] server.Server(415): Started @228309ms 2024-12-05T15:15:37,507 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:15:37,572 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:37,575 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:15:37,576 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:15:37,576 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:15:37,576 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:15:37,576 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@79eaf123{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:15:37,577 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@2899e0b7{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:15:37,691 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4d38e2e3{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/java.io.tmpdir/jetty-localhost-40405-hadoop-hdfs-3_4_1-tests_jar-_-any-1759613659479123576/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:37,691 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@5728e0aa{HTTP/1.1, (http/1.1)}{localhost:40405} 2024-12-05T15:15:37,691 INFO [Time-limited test {}] server.Server(415): Started @228507ms 2024-12-05T15:15:37,693 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:15:37,739 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:15:37,745 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:15:37,749 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:15:37,749 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:15:37,749 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:15:37,750 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3ac49840{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:15:37,750 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@5d858d58{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:15:37,795 WARN [Thread-1370 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data1/current/BP-1679678351-172.17.0.2-1733411737314/current, will proceed with Du for space computation calculation, 2024-12-05T15:15:37,796 WARN [Thread-1371 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data2/current/BP-1679678351-172.17.0.2-1733411737314/current, will proceed with Du for space computation calculation, 2024-12-05T15:15:37,820 WARN [Thread-1349 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:15:37,823 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6fa7248c15a98c6b with lease ID 0xd08d45b3e3c5c3a3: Processing first storage report for DS-d97451dd-c5ae-4bb0-a637-d097da243104 from datanode DatanodeRegistration(127.0.0.1:45861, datanodeUuid=06c36c51-7269-48cf-9467-b9873b0aa699, infoPort=37187, infoSecurePort=0, ipcPort=39077, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314) 2024-12-05T15:15:37,823 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6fa7248c15a98c6b with lease ID 0xd08d45b3e3c5c3a3: from storage DS-d97451dd-c5ae-4bb0-a637-d097da243104 node DatanodeRegistration(127.0.0.1:45861, datanodeUuid=06c36c51-7269-48cf-9467-b9873b0aa699, infoPort=37187, infoSecurePort=0, ipcPort=39077, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:37,823 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x6fa7248c15a98c6b with lease ID 0xd08d45b3e3c5c3a3: Processing first storage report for DS-f83bee46-d3c4-4a20-9e03-4fd285ab6448 from datanode DatanodeRegistration(127.0.0.1:45861, datanodeUuid=06c36c51-7269-48cf-9467-b9873b0aa699, infoPort=37187, infoSecurePort=0, ipcPort=39077, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314) 2024-12-05T15:15:37,823 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x6fa7248c15a98c6b with lease ID 0xd08d45b3e3c5c3a3: from storage DS-f83bee46-d3c4-4a20-9e03-4fd285ab6448 node DatanodeRegistration(127.0.0.1:45861, datanodeUuid=06c36c51-7269-48cf-9467-b9873b0aa699, infoPort=37187, infoSecurePort=0, ipcPort=39077, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:37,872 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@55215408{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/java.io.tmpdir/jetty-localhost-39639-hadoop-hdfs-3_4_1-tests_jar-_-any-4792583336075236556/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:15:37,872 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@262a1159{HTTP/1.1, (http/1.1)}{localhost:39639} 2024-12-05T15:15:37,872 INFO [Time-limited test {}] server.Server(415): Started @228688ms 2024-12-05T15:15:37,874 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:15:37,972 WARN [Thread-1396 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data3/current/BP-1679678351-172.17.0.2-1733411737314/current, will proceed with Du for space computation calculation, 2024-12-05T15:15:37,972 WARN [Thread-1397 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data4/current/BP-1679678351-172.17.0.2-1733411737314/current, will proceed with Du for space computation calculation, 2024-12-05T15:15:37,993 WARN [Thread-1385 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:15:37,996 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x65b83add4c784a85 with lease ID 0xd08d45b3e3c5c3a4: Processing first storage report for DS-349001f0-7e85-45f9-96f3-03129a4703fd from datanode DatanodeRegistration(127.0.0.1:37029, datanodeUuid=2a2cfa40-e6f8-4a02-98fb-1fc304ca18b9, infoPort=36425, infoSecurePort=0, ipcPort=37457, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314) 2024-12-05T15:15:37,996 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x65b83add4c784a85 with lease ID 0xd08d45b3e3c5c3a4: from storage DS-349001f0-7e85-45f9-96f3-03129a4703fd node DatanodeRegistration(127.0.0.1:37029, datanodeUuid=2a2cfa40-e6f8-4a02-98fb-1fc304ca18b9, infoPort=36425, infoSecurePort=0, ipcPort=37457, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:37,996 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x65b83add4c784a85 with lease ID 0xd08d45b3e3c5c3a4: Processing first storage report for DS-477580b9-cdc7-462a-9955-6a8889aee5c5 from datanode DatanodeRegistration(127.0.0.1:37029, datanodeUuid=2a2cfa40-e6f8-4a02-98fb-1fc304ca18b9, infoPort=36425, infoSecurePort=0, ipcPort=37457, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314) 2024-12-05T15:15:37,996 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x65b83add4c784a85 with lease ID 0xd08d45b3e3c5c3a4: from storage DS-477580b9-cdc7-462a-9955-6a8889aee5c5 node DatanodeRegistration(127.0.0.1:37029, datanodeUuid=2a2cfa40-e6f8-4a02-98fb-1fc304ca18b9, infoPort=36425, infoSecurePort=0, ipcPort=37457, storageInfo=lv=-57;cid=testClusterID;nsid=1081281460;c=1733411737314), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:15:37,999 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693 2024-12-05T15:15:38,005 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/zookeeper_0, clientPort=63920, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T15:15:38,006 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=63920 2024-12-05T15:15:38,006 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:38,008 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:38,010 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,010 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,010 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,011 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,011 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,011 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,016 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,016 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,017 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,019 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,019 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:15:38,020 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:15:38,022 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4 with version=8 2024-12-05T15:15:38,022 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/hbase-staging 2024-12-05T15:15:38,025 INFO [Time-limited test {}] client.ConnectionUtils(129): master/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:15:38,025 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:15:38,025 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:15:38,026 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:15:38,026 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:15:38,026 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,026 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:15:38,026 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:15:38,026 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:15:38,026 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,026 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,027 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,027 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:39961 2024-12-05T15:15:38,027 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:38,030 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:38,033 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:39961 connecting to ZooKeeper ensemble=127.0.0.1:63920 2024-12-05T15:15:38,039 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:399610x0, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:15:38,039 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:39961-0x10062bfca7b0000 connected 2024-12-05T15:15:38,064 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:15:38,065 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:15:38,065 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:15:38,068 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39961 2024-12-05T15:15:38,071 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39961 2024-12-05T15:15:38,072 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39961 2024-12-05T15:15:38,072 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39961 2024-12-05T15:15:38,073 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39961 2024-12-05T15:15:38,074 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4, hbase.cluster.distributed=false 2024-12-05T15:15:38,099 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:15:38,099 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:15:38,099 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:15:38,099 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:15:38,099 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:15:38,099 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:15:38,100 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:15:38,100 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:15:38,101 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:33595 2024-12-05T15:15:38,101 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T15:15:38,104 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T15:15:38,105 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:38,107 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:38,111 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:33595 connecting to ZooKeeper ensemble=127.0.0.1:63920 2024-12-05T15:15:38,114 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:335950x0, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:15:38,114 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:33595-0x10062bfca7b0001 connected 2024-12-05T15:15:38,114 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:15:38,115 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:15:38,115 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:15:38,116 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=33595 2024-12-05T15:15:38,116 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=33595 2024-12-05T15:15:38,116 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=33595 2024-12-05T15:15:38,117 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=33595 2024-12-05T15:15:38,117 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=33595 2024-12-05T15:15:38,117 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:38,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:15:38,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:15:38,120 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:38,122 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:15:38,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:15:38,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,123 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:15:38,123 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/4ecb0f57b706,39961,1733411738024 from backup master directory 2024-12-05T15:15:38,124 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:15:38,125 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:38,125 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:15:38,125 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:15:38,125 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:15:38,125 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:38,131 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;4ecb0f57b706:39961 2024-12-05T15:15:38,131 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:38,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:15:38,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:15:38,137 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/hbase.id with ID: 78c84c94-423a-423a-9572-0bed3a1a2e1d 2024-12-05T15:15:38,149 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:38,153 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,153 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,161 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:15:38,161 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:15:38,162 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:15:38,163 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T15:15:38,163 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:15:38,170 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:15:38,170 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:15:38,171 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store 2024-12-05T15:15:38,178 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:15:38,178 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:15:38,179 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:38,179 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:15:38,179 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:38,179 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:38,179 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:15:38,179 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:38,179 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:15:38,179 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:15:38,180 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/.initializing 2024-12-05T15:15:38,180 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/WALs/4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:38,183 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C39961%2C1733411738024, suffix=, logDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/WALs/4ecb0f57b706,39961,1733411738024, archiveDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/oldWALs, maxLogs=10 2024-12-05T15:15:38,184 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C39961%2C1733411738024.1733411738184 2024-12-05T15:15:38,193 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/WALs/4ecb0f57b706,39961,1733411738024/4ecb0f57b706%2C39961%2C1733411738024.1733411738184 2024-12-05T15:15:38,193 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37187:37187),(127.0.0.1/127.0.0.1:36425:36425)] 2024-12-05T15:15:38,193 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:15:38,193 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:38,193 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,193 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,195 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,196 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T15:15:38,196 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,197 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:15:38,197 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,198 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T15:15:38,198 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,198 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:15:38,198 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,199 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T15:15:38,199 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,200 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:15:38,200 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,201 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T15:15:38,201 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,202 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:15:38,203 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,203 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,206 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T15:15:38,207 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:15:38,209 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:15:38,210 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=777408, jitterRate=-0.011475786566734314}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T15:15:38,211 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:15:38,211 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T15:15:38,215 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@e5be6d5, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:15:38,216 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-05T15:15:38,216 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T15:15:38,216 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T15:15:38,216 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T15:15:38,217 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-05T15:15:38,217 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-05T15:15:38,217 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T15:15:38,219 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T15:15:38,220 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T15:15:38,222 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-05T15:15:38,222 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T15:15:38,223 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T15:15:38,224 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-05T15:15:38,224 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T15:15:38,225 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T15:15:38,227 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-05T15:15:38,227 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T15:15:38,228 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T15:15:38,230 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T15:15:38,231 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T15:15:38,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:15:38,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:15:38,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,233 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,235 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=4ecb0f57b706,39961,1733411738024, sessionid=0x10062bfca7b0000, setting cluster-up flag (Was=false) 2024-12-05T15:15:38,238 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,238 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,244 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T15:15:38,245 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:38,248 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,248 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,253 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T15:15:38,254 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:38,257 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-05T15:15:38,257 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-05T15:15:38,257 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 4ecb0f57b706,39961,1733411738024 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/4ecb0f57b706:0, corePoolSize=10, maxPoolSize=10 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:15:38,258 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733411768263 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T15:15:38,263 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,264 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T15:15:38,264 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:15:38,264 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T15:15:38,264 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-05T15:15:38,264 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T15:15:38,265 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,265 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:15:38,271 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T15:15:38,271 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T15:15:38,275 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411738271,5,FailOnTimeoutGroup] 2024-12-05T15:15:38,275 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411738275,5,FailOnTimeoutGroup] 2024-12-05T15:15:38,275 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,275 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T15:15:38,275 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,275 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,280 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:15:38,281 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:15:38,282 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-05T15:15:38,282 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4 2024-12-05T15:15:38,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:15:38,293 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:15:38,293 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:38,295 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:15:38,296 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:15:38,296 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,297 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:15:38,297 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:15:38,298 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:15:38,298 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,299 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:15:38,299 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:15:38,300 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:15:38,300 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,301 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:15:38,301 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740 2024-12-05T15:15:38,302 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740 2024-12-05T15:15:38,303 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:15:38,304 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:15:38,306 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:38,315 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:15:38,315 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=734803, jitterRate=-0.06564991176128387}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:15:38,316 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:15:38,316 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:15:38,316 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:15:38,316 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:15:38,316 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:15:38,316 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:15:38,316 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:15:38,316 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:15:38,318 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:15:38,318 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-05T15:15:38,318 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T15:15:38,319 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T15:15:38,321 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T15:15:38,344 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;4ecb0f57b706:33595 2024-12-05T15:15:38,345 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1008): ClusterId : 78c84c94-423a-423a-9572-0bed3a1a2e1d 2024-12-05T15:15:38,345 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T15:15:38,347 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T15:15:38,347 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T15:15:38,349 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T15:15:38,350 DEBUG [RS:0;4ecb0f57b706:33595 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@6c94cdf5, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:15:38,350 DEBUG [RS:0;4ecb0f57b706:33595 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@55fbefb3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:15:38,350 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-05T15:15:38,350 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-05T15:15:38,350 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-05T15:15:38,351 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,39961,1733411738024 with isa=4ecb0f57b706/172.17.0.2:33595, startcode=1733411738098 2024-12-05T15:15:38,351 DEBUG [RS:0;4ecb0f57b706:33595 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:15:38,353 INFO [RS-EventLoopGroup-10-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:41529, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:15:38,354 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39961 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,354 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39961 {}] master.ServerManager(486): Registering regionserver=4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,355 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4 2024-12-05T15:15:38,355 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:38195 2024-12-05T15:15:38,355 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-05T15:15:38,357 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:15:38,358 DEBUG [RS:0;4ecb0f57b706:33595 {}] zookeeper.ZKUtil(111): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,358 WARN [RS:0;4ecb0f57b706:33595 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:15:38,358 INFO [RS:0;4ecb0f57b706:33595 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:15:38,358 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,358 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [4ecb0f57b706,33595,1733411738098] 2024-12-05T15:15:38,362 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-05T15:15:38,362 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T15:15:38,364 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T15:15:38,367 INFO [RS:0;4ecb0f57b706:33595 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:15:38,367 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,367 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-05T15:15:38,368 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,368 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,369 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:15:38,369 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:15:38,369 DEBUG [RS:0;4ecb0f57b706:33595 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:15:38,370 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,370 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,370 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,370 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,370 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,33595,1733411738098-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:15:38,385 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T15:15:38,385 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,33595,1733411738098-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,400 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.Replication(204): 4ecb0f57b706,33595,1733411738098 started 2024-12-05T15:15:38,400 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1767): Serving as 4ecb0f57b706,33595,1733411738098, RpcServer on 4ecb0f57b706/172.17.0.2:33595, sessionid=0x10062bfca7b0001 2024-12-05T15:15:38,400 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T15:15:38,400 DEBUG [RS:0;4ecb0f57b706:33595 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,400 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,33595,1733411738098' 2024-12-05T15:15:38,400 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T15:15:38,401 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T15:15:38,401 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T15:15:38,401 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T15:15:38,401 DEBUG [RS:0;4ecb0f57b706:33595 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,401 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,33595,1733411738098' 2024-12-05T15:15:38,401 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T15:15:38,402 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T15:15:38,402 DEBUG [RS:0;4ecb0f57b706:33595 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T15:15:38,402 INFO [RS:0;4ecb0f57b706:33595 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T15:15:38,402 INFO [RS:0;4ecb0f57b706:33595 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T15:15:38,471 WARN [4ecb0f57b706:39961 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-05T15:15:38,504 INFO [RS:0;4ecb0f57b706:33595 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C33595%2C1733411738098, suffix=, logDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098, archiveDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/oldWALs, maxLogs=32 2024-12-05T15:15:38,505 INFO [RS:0;4ecb0f57b706:33595 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C33595%2C1733411738098.1733411738505 2024-12-05T15:15:38,513 INFO [RS:0;4ecb0f57b706:33595 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411738505 2024-12-05T15:15:38,513 DEBUG [RS:0;4ecb0f57b706:33595 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37187:37187),(127.0.0.1/127.0.0.1:36425:36425)] 2024-12-05T15:15:38,528 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:15:38,529 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,529 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,529 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,530 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,545 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,546 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,546 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,546 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,546 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,547 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,548 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,548 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,549 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,551 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:38,721 DEBUG [4ecb0f57b706:39961 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-05T15:15:38,721 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,722 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,33595,1733411738098, state=OPENING 2024-12-05T15:15:38,724 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T15:15:38,726 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,726 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:38,726 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=4ecb0f57b706,33595,1733411738098}] 2024-12-05T15:15:38,726 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:15:38,726 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:15:38,879 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,879 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T15:15:38,881 INFO [RS-EventLoopGroup-11-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37342, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T15:15:38,885 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-05T15:15:38,885 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:15:38,887 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C33595%2C1733411738098.meta, suffix=.meta, logDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098, archiveDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/oldWALs, maxLogs=32 2024-12-05T15:15:38,887 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C33595%2C1733411738098.meta.1733411738887.meta 2024-12-05T15:15:38,892 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.meta.1733411738887.meta 2024-12-05T15:15:38,892 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36425:36425),(127.0.0.1/127.0.0.1:37187:37187)] 2024-12-05T15:15:38,892 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:15:38,893 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T15:15:38,893 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T15:15:38,893 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T15:15:38,893 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T15:15:38,893 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:38,893 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-05T15:15:38,893 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-05T15:15:38,894 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:15:38,895 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:15:38,895 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,895 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:15:38,896 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:15:38,896 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:15:38,896 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,897 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:15:38,897 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:15:38,897 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:15:38,897 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,898 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:15:38,899 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740 2024-12-05T15:15:38,899 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740 2024-12-05T15:15:38,901 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:15:38,902 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:15:38,903 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=693275, jitterRate=-0.11845535039901733}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:15:38,903 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:15:38,904 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733411738879 2024-12-05T15:15:38,906 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T15:15:38,906 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-05T15:15:38,906 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:38,907 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,33595,1733411738098, state=OPEN 2024-12-05T15:15:38,912 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:15:38,912 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:15:38,912 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:15:38,912 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:15:38,914 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T15:15:38,914 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=4ecb0f57b706,33595,1733411738098 in 186 msec 2024-12-05T15:15:38,916 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T15:15:38,916 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 596 msec 2024-12-05T15:15:38,918 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 660 msec 2024-12-05T15:15:38,918 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733411738918, completionTime=-1 2024-12-05T15:15:38,918 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-05T15:15:38,918 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-05T15:15:38,919 DEBUG [hconnection-0x7da3e28-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:15:38,920 INFO [RS-EventLoopGroup-11-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37344, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:15:38,921 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-05T15:15:38,921 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733411798921 2024-12-05T15:15:38,921 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733411858921 2024-12-05T15:15:38,921 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 2 msec 2024-12-05T15:15:38,926 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39961,1733411738024-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,926 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39961,1733411738024-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,926 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39961,1733411738024-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,926 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-4ecb0f57b706:39961, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,926 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:38,926 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-05T15:15:38,926 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:15:38,928 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-05T15:15:38,928 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-05T15:15:38,929 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:15:38,929 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:38,930 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:15:38,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:15:38,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:15:38,939 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 8dd9850309053e109fcaf424b37afb54, NAME => 'hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4 2024-12-05T15:15:38,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:15:38,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:15:38,948 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:38,948 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 8dd9850309053e109fcaf424b37afb54, disabling compactions & flushes 2024-12-05T15:15:38,948 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:38,948 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:38,948 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. after waiting 0 ms 2024-12-05T15:15:38,948 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:38,948 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:38,948 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 8dd9850309053e109fcaf424b37afb54: 2024-12-05T15:15:38,949 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:15:38,949 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733411738949"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411738949"}]},"ts":"1733411738949"} 2024-12-05T15:15:38,951 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:15:38,952 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:15:38,952 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411738952"}]},"ts":"1733411738952"} 2024-12-05T15:15:38,954 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-05T15:15:38,958 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=8dd9850309053e109fcaf424b37afb54, ASSIGN}] 2024-12-05T15:15:38,959 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=8dd9850309053e109fcaf424b37afb54, ASSIGN 2024-12-05T15:15:38,960 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=8dd9850309053e109fcaf424b37afb54, ASSIGN; state=OFFLINE, location=4ecb0f57b706,33595,1733411738098; forceNewPlan=false, retain=false 2024-12-05T15:15:39,111 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=8dd9850309053e109fcaf424b37afb54, regionState=OPENING, regionLocation=4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:39,113 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 8dd9850309053e109fcaf424b37afb54, server=4ecb0f57b706,33595,1733411738098}] 2024-12-05T15:15:39,132 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:39,266 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:39,270 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:39,270 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 8dd9850309053e109fcaf424b37afb54, NAME => 'hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:15:39,270 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 8dd9850309053e109fcaf424b37afb54 2024-12-05T15:15:39,270 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:39,270 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 8dd9850309053e109fcaf424b37afb54 2024-12-05T15:15:39,270 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 8dd9850309053e109fcaf424b37afb54 2024-12-05T15:15:39,272 INFO [StoreOpener-8dd9850309053e109fcaf424b37afb54-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 8dd9850309053e109fcaf424b37afb54 2024-12-05T15:15:39,274 INFO [StoreOpener-8dd9850309053e109fcaf424b37afb54-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 8dd9850309053e109fcaf424b37afb54 columnFamilyName info 2024-12-05T15:15:39,274 DEBUG [StoreOpener-8dd9850309053e109fcaf424b37afb54-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:39,274 INFO [StoreOpener-8dd9850309053e109fcaf424b37afb54-1 {}] regionserver.HStore(327): Store=8dd9850309053e109fcaf424b37afb54/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:15:39,275 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54 2024-12-05T15:15:39,276 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54 2024-12-05T15:15:39,277 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 8dd9850309053e109fcaf424b37afb54 2024-12-05T15:15:39,279 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:15:39,280 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 8dd9850309053e109fcaf424b37afb54; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=728608, jitterRate=-0.07352818548679352}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:15:39,280 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 8dd9850309053e109fcaf424b37afb54: 2024-12-05T15:15:39,281 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54., pid=6, masterSystemTime=1733411739266 2024-12-05T15:15:39,283 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:39,283 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:39,283 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=8dd9850309053e109fcaf424b37afb54, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:39,287 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T15:15:39,287 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 8dd9850309053e109fcaf424b37afb54, server=4ecb0f57b706,33595,1733411738098 in 172 msec 2024-12-05T15:15:39,289 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T15:15:39,289 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=8dd9850309053e109fcaf424b37afb54, ASSIGN in 329 msec 2024-12-05T15:15:39,290 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:15:39,290 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411739290"}]},"ts":"1733411739290"} 2024-12-05T15:15:39,291 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-05T15:15:39,294 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:15:39,295 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 367 msec 2024-12-05T15:15:39,307 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:39,329 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-05T15:15:39,331 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:15:39,331 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:39,331 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:15:39,336 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-05T15:15:39,344 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:15:39,347 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 11 msec 2024-12-05T15:15:39,358 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-05T15:15:39,365 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:15:39,369 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 10 msec 2024-12-05T15:15:39,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-05T15:15:39,386 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.261sec 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39961,1733411738024-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:15:39,387 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39961,1733411738024-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T15:15:39,389 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-05T15:15:39,389 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T15:15:39,389 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39961,1733411738024-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:15:39,428 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x62253a67 to 127.0.0.1:63920 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@23ae7e64 2024-12-05T15:15:39,431 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3b4b904e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:15:39,433 DEBUG [hconnection-0x48290c7a-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:15:39,434 INFO [RS-EventLoopGroup-11-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:37348, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:15:39,436 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=4ecb0f57b706,39961,1733411738024 2024-12-05T15:15:39,436 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:15:39,439 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-05T15:15:39,440 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-05T15:15:39,442 INFO [RS-EventLoopGroup-10-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:54388, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-05T15:15:39,443 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-05T15:15:39,443 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-05T15:15:39,443 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testCompactionRecordDoesntBlockRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:15:39,444 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:15:39,445 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:15:39,445 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:39,445 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testCompactionRecordDoesntBlockRolling" procId is: 9 2024-12-05T15:15:39,446 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:15:39,446 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:15:39,453 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741837_1013 (size=405) 2024-12-05T15:15:39,454 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741837_1013 (size=405) 2024-12-05T15:15:39,455 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 26be18c713c8da53cab1008ab6b5dbbc, NAME => 'TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testCompactionRecordDoesntBlockRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4 2024-12-05T15:15:39,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741838_1014 (size=88) 2024-12-05T15:15:39,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741838_1014 (size=88) 2024-12-05T15:15:39,462 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:39,462 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1681): Closing 26be18c713c8da53cab1008ab6b5dbbc, disabling compactions & flushes 2024-12-05T15:15:39,462 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:39,462 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:39,462 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. after waiting 0 ms 2024-12-05T15:15:39,462 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:39,462 INFO [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:39,462 DEBUG [RegionOpenAndInit-TestLogRolling-testCompactionRecordDoesntBlockRolling-pool-0 {}] regionserver.HRegion(1635): Region close journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:15:39,464 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:15:39,464 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.","families":{"info":[{"qualifier":"regioninfo","vlen":87,"tag":[],"timestamp":"1733411739464"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411739464"}]},"ts":"1733411739464"} 2024-12-05T15:15:39,465 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:15:39,466 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:15:39,466 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411739466"}]},"ts":"1733411739466"} 2024-12-05T15:15:39,468 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testCompactionRecordDoesntBlockRolling, state=ENABLING in hbase:meta 2024-12-05T15:15:39,471 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=26be18c713c8da53cab1008ab6b5dbbc, ASSIGN}] 2024-12-05T15:15:39,472 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=26be18c713c8da53cab1008ab6b5dbbc, ASSIGN 2024-12-05T15:15:39,473 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=26be18c713c8da53cab1008ab6b5dbbc, ASSIGN; state=OFFLINE, location=4ecb0f57b706,33595,1733411738098; forceNewPlan=false, retain=false 2024-12-05T15:15:39,624 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=26be18c713c8da53cab1008ab6b5dbbc, regionState=OPENING, regionLocation=4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:39,626 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 26be18c713c8da53cab1008ab6b5dbbc, server=4ecb0f57b706,33595,1733411738098}] 2024-12-05T15:15:39,778 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:39,782 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:39,782 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 26be18c713c8da53cab1008ab6b5dbbc, NAME => 'TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:15:39,782 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testCompactionRecordDoesntBlockRolling 26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:15:39,783 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:15:39,783 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:15:39,783 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:15:39,784 INFO [StoreOpener-26be18c713c8da53cab1008ab6b5dbbc-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:15:39,785 INFO [StoreOpener-26be18c713c8da53cab1008ab6b5dbbc-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 26be18c713c8da53cab1008ab6b5dbbc columnFamilyName info 2024-12-05T15:15:39,785 DEBUG [StoreOpener-26be18c713c8da53cab1008ab6b5dbbc-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:15:39,786 INFO [StoreOpener-26be18c713c8da53cab1008ab6b5dbbc-1 {}] regionserver.HStore(327): Store=26be18c713c8da53cab1008ab6b5dbbc/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:15:39,787 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:15:39,787 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:15:39,789 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:15:39,791 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:15:39,791 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened 26be18c713c8da53cab1008ab6b5dbbc; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=805837, jitterRate=0.024675488471984863}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:15:39,792 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:15:39,793 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc., pid=11, masterSystemTime=1733411739778 2024-12-05T15:15:39,795 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:39,795 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:39,795 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=26be18c713c8da53cab1008ab6b5dbbc, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:39,799 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-05T15:15:39,799 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 26be18c713c8da53cab1008ab6b5dbbc, server=4ecb0f57b706,33595,1733411738098 in 171 msec 2024-12-05T15:15:39,801 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-05T15:15:39,801 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling, region=26be18c713c8da53cab1008ab6b5dbbc, ASSIGN in 328 msec 2024-12-05T15:15:39,802 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:15:39,802 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testCompactionRecordDoesntBlockRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411739802"}]},"ts":"1733411739802"} 2024-12-05T15:15:39,804 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testCompactionRecordDoesntBlockRolling, state=ENABLED in hbase:meta 2024-12-05T15:15:39,807 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:15:39,808 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 364 msec 2024-12-05T15:15:40,132 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:40,308 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:41,133 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:41,308 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:42,134 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:42,309 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:43,134 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:43,309 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:43,310 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=2 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta after 68044ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor237.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:15:44,135 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:44,310 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:44,395 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:15:44,396 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,396 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,397 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,397 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,412 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,412 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,412 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,412 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,412 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,413 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,416 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,416 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,416 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,419 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:15:44,431 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testCompactionRecordDoesntBlockRolling' 2024-12-05T15:15:44,990 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:15:44,991 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling Metrics about Tables on a single HBase RegionServer 2024-12-05T15:15:45,136 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:45,311 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:46,136 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:46,311 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:47,137 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:47,312 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:48,137 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:48,312 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:49,138 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:49,313 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:49,448 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:15:49,448 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 9 completed 2024-12-05T15:15:49,450 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:15:49,450 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:49,458 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush hbase:namespace 2024-12-05T15:15:49,463 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] procedure2.ProcedureExecutor(1098): Stored pid=12, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace 2024-12-05T15:15:49,464 INFO [PEWorker-4 {}] procedure.FlushTableProcedure(91): pid=12, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace execute state=FLUSH_TABLE_PREPARE 2024-12-05T15:15:49,464 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2024-12-05T15:15:49,465 INFO [PEWorker-4 {}] procedure.FlushTableProcedure(91): pid=12, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-05T15:15:49,467 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=13, ppid=12, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-05T15:15:49,626 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:49,627 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=33595 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=13 2024-12-05T15:15:49,627 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.FlushRegionCallable(51): Starting region operation on hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:49,628 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegion(2837): Flushing 8dd9850309053e109fcaf424b37afb54 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-05T15:15:49,645 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54/.tmp/info/fdb4e2f87b774a77858f0d4fa7e89cca is 45, key is default/info:d/1733411739340/Put/seqid=0 2024-12-05T15:15:49,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741839_1015 (size=5037) 2024-12-05T15:15:49,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741839_1015 (size=5037) 2024-12-05T15:15:49,650 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54/.tmp/info/fdb4e2f87b774a77858f0d4fa7e89cca 2024-12-05T15:15:49,658 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54/.tmp/info/fdb4e2f87b774a77858f0d4fa7e89cca as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54/info/fdb4e2f87b774a77858f0d4fa7e89cca 2024-12-05T15:15:49,663 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54/info/fdb4e2f87b774a77858f0d4fa7e89cca, entries=2, sequenceid=6, filesize=4.9 K 2024-12-05T15:15:49,664 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 8dd9850309053e109fcaf424b37afb54 in 37ms, sequenceid=6, compaction requested=false 2024-12-05T15:15:49,664 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.HRegion(2538): Flush status journal for 8dd9850309053e109fcaf424b37afb54: 2024-12-05T15:15:49,664 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.FlushRegionCallable(64): Closing region operation on hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:15:49,666 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=13}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=13 2024-12-05T15:15:49,667 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster(4106): Remote procedure done, pid=13 2024-12-05T15:15:49,670 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=13, resume processing ppid=12 2024-12-05T15:15:49,670 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=13, ppid=12, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 201 msec 2024-12-05T15:15:49,672 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=12, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=12, table=hbase:namespace in 212 msec 2024-12-05T15:15:50,139 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:50,314 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:51,139 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:51,314 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:52,140 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:52,315 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:53,140 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:53,315 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:54,141 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:54,316 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:55,141 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:55,317 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:56,142 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:56,317 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:57,143 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:57,318 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:58,143 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:58,318 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:59,144 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:59,319 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:15:59,466 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2024-12-05T15:15:59,500 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: hbase:namespace, procId: 12 completed 2024-12-05T15:15:59,507 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:15:59,508 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] procedure2.ProcedureExecutor(1098): Stored pid=14, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:15:59,509 INFO [PEWorker-3 {}] procedure.FlushTableProcedure(91): pid=14, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-05T15:15:59,509 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=14 2024-12-05T15:15:59,509 INFO [PEWorker-3 {}] procedure.FlushTableProcedure(91): pid=14, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-05T15:15:59,510 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=15, ppid=14, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-05T15:15:59,662 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:15:59,662 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=33595 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=15 2024-12-05T15:15:59,663 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:59,663 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegion(2837): Flushing 26be18c713c8da53cab1008ab6b5dbbc 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-05T15:15:59,680 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/3e70d4efe9924928826be8a9b845500a is 1080, key is row0001/info:/1733411759503/Put/seqid=0 2024-12-05T15:15:59,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741840_1016 (size=6033) 2024-12-05T15:15:59,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741840_1016 (size=6033) 2024-12-05T15:15:59,686 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/3e70d4efe9924928826be8a9b845500a 2024-12-05T15:15:59,692 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/3e70d4efe9924928826be8a9b845500a as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/3e70d4efe9924928826be8a9b845500a 2024-12-05T15:15:59,698 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/3e70d4efe9924928826be8a9b845500a, entries=1, sequenceid=5, filesize=5.9 K 2024-12-05T15:15:59,699 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 26be18c713c8da53cab1008ab6b5dbbc in 35ms, sequenceid=5, compaction requested=false 2024-12-05T15:15:59,699 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.HRegion(2538): Flush status journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:15:59,699 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:15:59,699 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=15}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=15 2024-12-05T15:15:59,699 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster(4106): Remote procedure done, pid=15 2024-12-05T15:15:59,702 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=15, resume processing ppid=14 2024-12-05T15:15:59,702 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=15, ppid=14, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 190 msec 2024-12-05T15:15:59,704 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=14, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=14, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 196 msec 2024-12-05T15:16:00,144 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:00,319 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:01,145 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:01,320 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:02,145 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:02,321 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:03,146 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:03,321 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:03,882 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 817a20dfbf0d1531445f95325bcaf27c, had cached 0 bytes from a total of 23930 2024-12-05T15:16:04,146 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:04,322 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:05,147 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:05,322 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:06,148 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:06,323 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:07,148 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:07,323 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:07,999 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:16:08,149 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:08,324 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:09,149 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:09,325 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:09,510 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=14 2024-12-05T15:16:09,510 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 14 completed 2024-12-05T15:16:09,515 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:16:09,516 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] procedure2.ProcedureExecutor(1098): Stored pid=16, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:16:09,516 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=16 2024-12-05T15:16:09,517 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=16, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-05T15:16:09,517 INFO [PEWorker-2 {}] procedure.FlushTableProcedure(91): pid=16, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-05T15:16:09,518 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=17, ppid=16, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-05T15:16:09,669 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:16:09,670 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=33595 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=17 2024-12-05T15:16:09,670 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:09,671 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegion(2837): Flushing 26be18c713c8da53cab1008ab6b5dbbc 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-05T15:16:09,675 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/c725144f29074ffab03d5aa502ec88d8 is 1080, key is row0002/info:/1733411769511/Put/seqid=0 2024-12-05T15:16:09,680 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741841_1017 (size=6033) 2024-12-05T15:16:09,681 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741841_1017 (size=6033) 2024-12-05T15:16:09,681 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/c725144f29074ffab03d5aa502ec88d8 2024-12-05T15:16:09,687 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/c725144f29074ffab03d5aa502ec88d8 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/c725144f29074ffab03d5aa502ec88d8 2024-12-05T15:16:09,691 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/c725144f29074ffab03d5aa502ec88d8, entries=1, sequenceid=9, filesize=5.9 K 2024-12-05T15:16:09,692 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 26be18c713c8da53cab1008ab6b5dbbc in 21ms, sequenceid=9, compaction requested=false 2024-12-05T15:16:09,693 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.HRegion(2538): Flush status journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:16:09,693 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:09,693 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-2 {event_type=RS_FLUSH_REGIONS, pid=17}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=17 2024-12-05T15:16:09,693 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster(4106): Remote procedure done, pid=17 2024-12-05T15:16:09,696 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=17, resume processing ppid=16 2024-12-05T15:16:09,696 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=17, ppid=16, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 176 msec 2024-12-05T15:16:09,697 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=16, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=16, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 181 msec 2024-12-05T15:16:10,150 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:10,325 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:10,350 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:16:10,352 INFO [RS-EventLoopGroup-10-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:32868, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.4 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:16:11,150 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:11,326 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:12,151 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:12,326 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:13,152 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:13,327 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:14,152 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:14,328 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:15,153 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:15,328 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:16,153 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:16,329 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:17,154 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:17,329 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:18,155 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:18,330 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:19,155 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:19,331 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:19,518 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=16 2024-12-05T15:16:19,518 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 16 completed 2024-12-05T15:16:19,520 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C33595%2C1733411738098.1733411779520 2024-12-05T15:16:19,527 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411738505 with entries=13, filesize=6.41 KB; new WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411779520 2024-12-05T15:16:19,528 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37187:37187),(127.0.0.1/127.0.0.1:36425:36425)] 2024-12-05T15:16:19,528 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411738505 is not closed yet, will try archiving it next time 2024-12-05T15:16:19,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741833_1009 (size=6574) 2024-12-05T15:16:19,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741833_1009 (size=6574) 2024-12-05T15:16:19,531 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:16:19,532 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] procedure2.ProcedureExecutor(1098): Stored pid=18, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:16:19,533 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=18 2024-12-05T15:16:19,533 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=18, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-05T15:16:19,534 INFO [PEWorker-1 {}] procedure.FlushTableProcedure(91): pid=18, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-05T15:16:19,534 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=19, ppid=18, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-05T15:16:19,611 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-05T15:16:19,611 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-05T15:16:19,686 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:16:19,687 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=33595 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=19 2024-12-05T15:16:19,687 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:19,687 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegion(2837): Flushing 26be18c713c8da53cab1008ab6b5dbbc 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-05T15:16:19,691 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/bf704d6b87f9409b852e4f0bda67f40a is 1080, key is row0003/info:/1733411779519/Put/seqid=0 2024-12-05T15:16:19,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741843_1019 (size=6033) 2024-12-05T15:16:19,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741843_1019 (size=6033) 2024-12-05T15:16:19,698 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=13 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/bf704d6b87f9409b852e4f0bda67f40a 2024-12-05T15:16:19,705 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/bf704d6b87f9409b852e4f0bda67f40a as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/bf704d6b87f9409b852e4f0bda67f40a 2024-12-05T15:16:19,710 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/bf704d6b87f9409b852e4f0bda67f40a, entries=1, sequenceid=13, filesize=5.9 K 2024-12-05T15:16:19,711 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 26be18c713c8da53cab1008ab6b5dbbc in 23ms, sequenceid=13, compaction requested=true 2024-12-05T15:16:19,711 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.HRegion(2538): Flush status journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:16:19,711 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:19,711 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-0 {event_type=RS_FLUSH_REGIONS, pid=19}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=19 2024-12-05T15:16:19,711 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster(4106): Remote procedure done, pid=19 2024-12-05T15:16:19,714 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=19, resume processing ppid=18 2024-12-05T15:16:19,714 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=19, ppid=18, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 178 msec 2024-12-05T15:16:19,716 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=18, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=18, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 184 msec 2024-12-05T15:16:20,156 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:20,331 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:21,156 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:21,332 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:22,157 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:22,332 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:23,157 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:23,333 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:24,158 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:24,270 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 8dd9850309053e109fcaf424b37afb54, had cached 0 bytes from a total of 5037 2024-12-05T15:16:24,333 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:24,783 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 26be18c713c8da53cab1008ab6b5dbbc, had cached 0 bytes from a total of 18099 2024-12-05T15:16:25,159 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:25,334 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:26,159 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:26,334 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:27,160 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:27,335 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:28,160 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:28,336 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:29,161 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:29,336 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:29,534 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=18 2024-12-05T15:16:29,535 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 18 completed 2024-12-05T15:16:29,535 DEBUG [Time-limited test {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:16:29,536 DEBUG [Time-limited test {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 18099 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:16:29,536 DEBUG [Time-limited test {}] regionserver.HStore(1540): 26be18c713c8da53cab1008ab6b5dbbc/info is initiating minor compaction (all files) 2024-12-05T15:16:29,536 INFO [Time-limited test {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:16:29,537 INFO [Time-limited test {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:29,537 INFO [Time-limited test {}] regionserver.HRegion(2351): Starting compaction of 26be18c713c8da53cab1008ab6b5dbbc/info in TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:29,537 INFO [Time-limited test {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/3e70d4efe9924928826be8a9b845500a, hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/c725144f29074ffab03d5aa502ec88d8, hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/bf704d6b87f9409b852e4f0bda67f40a] into tmpdir=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp, totalSize=17.7 K 2024-12-05T15:16:29,538 DEBUG [Time-limited test {}] compactions.Compactor(224): Compacting 3e70d4efe9924928826be8a9b845500a, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=5, earliestPutTs=1733411759503 2024-12-05T15:16:29,538 DEBUG [Time-limited test {}] compactions.Compactor(224): Compacting c725144f29074ffab03d5aa502ec88d8, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=9, earliestPutTs=1733411769511 2024-12-05T15:16:29,538 DEBUG [Time-limited test {}] compactions.Compactor(224): Compacting bf704d6b87f9409b852e4f0bda67f40a, keycount=1, bloomtype=ROW, size=5.9 K, encoding=NONE, compression=NONE, seqNum=13, earliestPutTs=1733411779519 2024-12-05T15:16:29,549 INFO [Time-limited test {}] throttle.PressureAwareThroughputController(145): 26be18c713c8da53cab1008ab6b5dbbc#info#compaction#29 average throughput is unlimited, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:16:29,550 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/af799aed42d94a0ab08dfac7583bde7a is 1080, key is row0001/info:/1733411759503/Put/seqid=0 2024-12-05T15:16:29,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741844_1020 (size=8296) 2024-12-05T15:16:29,555 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741844_1020 (size=8296) 2024-12-05T15:16:29,561 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/af799aed42d94a0ab08dfac7583bde7a as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/af799aed42d94a0ab08dfac7583bde7a 2024-12-05T15:16:29,566 INFO [Time-limited test {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 26be18c713c8da53cab1008ab6b5dbbc/info of 26be18c713c8da53cab1008ab6b5dbbc into af799aed42d94a0ab08dfac7583bde7a(size=8.1 K), total size for store is 8.1 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:16:29,566 DEBUG [Time-limited test {}] regionserver.HRegion(2381): Compaction status journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:16:29,568 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C33595%2C1733411738098.1733411789568 2024-12-05T15:16:29,576 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411779520 with entries=4, filesize=2.45 KB; new WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411789568 2024-12-05T15:16:29,576 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36425:36425),(127.0.0.1/127.0.0.1:37187:37187)] 2024-12-05T15:16:29,576 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411779520 is not closed yet, will try archiving it next time 2024-12-05T15:16:29,576 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411738505 to hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/oldWALs/4ecb0f57b706%2C33595%2C1733411738098.1733411738505 2024-12-05T15:16:29,577 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741842_1018 (size=2520) 2024-12-05T15:16:29,578 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741842_1018 (size=2520) 2024-12-05T15:16:29,580 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster$22(4386): Client=jenkins//172.17.0.2 flush TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:16:29,581 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] procedure2.ProcedureExecutor(1098): Stored pid=20, state=RUNNABLE:FLUSH_TABLE_PREPARE; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:16:29,582 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2024-12-05T15:16:29,582 INFO [PEWorker-5 {}] procedure.FlushTableProcedure(91): pid=20, state=RUNNABLE:FLUSH_TABLE_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_PREPARE 2024-12-05T15:16:29,583 INFO [PEWorker-5 {}] procedure.FlushTableProcedure(91): pid=20, state=RUNNABLE:FLUSH_TABLE_FLUSH_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling execute state=FLUSH_TABLE_FLUSH_REGIONS 2024-12-05T15:16:29,583 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=21, ppid=20, state=RUNNABLE; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure}] 2024-12-05T15:16:29,735 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,33595,1733411738098 2024-12-05T15:16:29,735 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=33595 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.FlushRegionCallable, pid=21 2024-12-05T15:16:29,736 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.FlushRegionCallable(51): Starting region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:29,736 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegion(2837): Flushing 26be18c713c8da53cab1008ab6b5dbbc 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-05T15:16:29,740 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/91ab1e8cf9f3401e871aa81322b50548 is 1080, key is row0000/info:/1733411789567/Put/seqid=0 2024-12-05T15:16:29,746 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741846_1022 (size=6033) 2024-12-05T15:16:29,746 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741846_1022 (size=6033) 2024-12-05T15:16:29,746 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=18 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/91ab1e8cf9f3401e871aa81322b50548 2024-12-05T15:16:29,752 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/91ab1e8cf9f3401e871aa81322b50548 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/91ab1e8cf9f3401e871aa81322b50548 2024-12-05T15:16:29,757 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/91ab1e8cf9f3401e871aa81322b50548, entries=1, sequenceid=18, filesize=5.9 K 2024-12-05T15:16:29,758 INFO [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 26be18c713c8da53cab1008ab6b5dbbc in 22ms, sequenceid=18, compaction requested=false 2024-12-05T15:16:29,759 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.HRegion(2538): Flush status journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:16:29,759 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.FlushRegionCallable(64): Closing region operation on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:29,759 DEBUG [RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0-1 {event_type=RS_FLUSH_REGIONS, pid=21}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=21 2024-12-05T15:16:29,759 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.HMaster(4106): Remote procedure done, pid=21 2024-12-05T15:16:29,762 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=21, resume processing ppid=20 2024-12-05T15:16:29,762 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=21, ppid=20, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushRegionProcedure in 177 msec 2024-12-05T15:16:29,764 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=20, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.FlushTableProcedure, id=20, table=TestLogRolling-testCompactionRecordDoesntBlockRolling in 183 msec 2024-12-05T15:16:30,161 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:30,337 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:31,162 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:31,337 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:32,163 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:32,338 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:33,163 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:33,338 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:34,164 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:34,339 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:35,164 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:35,340 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:36,165 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:36,340 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:37,165 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:37,341 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:37,999 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:16:38,166 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:38,341 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:38,969 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 8dd9850309053e109fcaf424b37afb54 changed from -1.0 to 0.0, refreshing cache 2024-12-05T15:16:39,166 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:39,342 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:39,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39961 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2024-12-05T15:16:39,583 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: FLUSH, Table Name: default:TestLogRolling-testCompactionRecordDoesntBlockRolling, procId: 20 completed 2024-12-05T15:16:39,585 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C33595%2C1733411738098.1733411799585 2024-12-05T15:16:39,591 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411789568 with entries=3, filesize=1.97 KB; new WAL /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411799585 2024-12-05T15:16:39,592 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:36425:36425),(127.0.0.1/127.0.0.1:37187:37187)] 2024-12-05T15:16:39,592 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411789568 is not closed yet, will try archiving it next time 2024-12-05T15:16:39,592 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098/4ecb0f57b706%2C33595%2C1733411738098.1733411779520 to hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/oldWALs/4ecb0f57b706%2C33595%2C1733411738098.1733411779520 2024-12-05T15:16:39,592 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-05T15:16:39,592 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-05T15:16:39,592 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x62253a67 to 127.0.0.1:63920 2024-12-05T15:16:39,592 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:16:39,592 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T15:16:39,592 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=559026745, stopped=false 2024-12-05T15:16:39,592 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=4ecb0f57b706,39961,1733411738024 2024-12-05T15:16:39,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741845_1021 (size=2026) 2024-12-05T15:16:39,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741845_1021 (size=2026) 2024-12-05T15:16:39,594 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:16:39,594 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:16:39,594 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-05T15:16:39,594 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:39,594 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:39,594 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:16:39,594 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,33595,1733411738098' ***** 2024-12-05T15:16:39,594 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-05T15:16:39,595 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:16:39,595 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T15:16:39,595 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-05T15:16:39,595 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:16:39,595 INFO [RS:0;4ecb0f57b706:33595 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(3579): Received CLOSE for 8dd9850309053e109fcaf424b37afb54 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(3579): Received CLOSE for 26be18c713c8da53cab1008ab6b5dbbc 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,33595,1733411738098 2024-12-05T15:16:39,596 DEBUG [RS:0;4ecb0f57b706:33595 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:16:39,596 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 8dd9850309053e109fcaf424b37afb54, disabling compactions & flushes 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T15:16:39,596 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:16:39,596 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:16:39,596 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. after waiting 0 ms 2024-12-05T15:16:39,596 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:16:39,596 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1599): Waiting on 3 regions to close 2024-12-05T15:16:39,596 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1603): Online Regions={1588230740=hbase:meta,,1.1588230740, 8dd9850309053e109fcaf424b37afb54=hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54., 26be18c713c8da53cab1008ab6b5dbbc=TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.} 2024-12-05T15:16:39,596 DEBUG [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 26be18c713c8da53cab1008ab6b5dbbc, 8dd9850309053e109fcaf424b37afb54 2024-12-05T15:16:39,596 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:16:39,596 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:16:39,596 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:16:39,597 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:16:39,597 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:16:39,597 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=3.05 KB heapSize=5.55 KB 2024-12-05T15:16:39,600 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/namespace/8dd9850309053e109fcaf424b37afb54/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-05T15:16:39,601 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:16:39,601 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 8dd9850309053e109fcaf424b37afb54: 2024-12-05T15:16:39,601 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733411738926.8dd9850309053e109fcaf424b37afb54. 2024-12-05T15:16:39,601 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 26be18c713c8da53cab1008ab6b5dbbc, disabling compactions & flushes 2024-12-05T15:16:39,601 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:39,601 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:39,601 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. after waiting 0 ms 2024-12-05T15:16:39,601 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:39,601 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 26be18c713c8da53cab1008ab6b5dbbc 1/1 column families, dataSize=1.05 KB heapSize=1.38 KB 2024-12-05T15:16:39,604 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/968b951b83f9444e8c3695bdd5ab5792 is 1080, key is row0001/info:/1733411799584/Put/seqid=0 2024-12-05T15:16:39,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741848_1024 (size=6033) 2024-12-05T15:16:39,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741848_1024 (size=6033) 2024-12-05T15:16:39,615 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.05 KB at sequenceid=22 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/968b951b83f9444e8c3695bdd5ab5792 2024-12-05T15:16:39,619 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/.tmp/info/cc43432509cf46e8b42abdfd237c3a82 is 227, key is TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc./info:regioninfo/1733411739795/Put/seqid=0 2024-12-05T15:16:39,622 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/.tmp/info/968b951b83f9444e8c3695bdd5ab5792 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/968b951b83f9444e8c3695bdd5ab5792 2024-12-05T15:16:39,624 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741849_1025 (size=8430) 2024-12-05T15:16:39,624 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741849_1025 (size=8430) 2024-12-05T15:16:39,625 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.79 KB at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/.tmp/info/cc43432509cf46e8b42abdfd237c3a82 2024-12-05T15:16:39,628 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/968b951b83f9444e8c3695bdd5ab5792, entries=1, sequenceid=22, filesize=5.9 K 2024-12-05T15:16:39,629 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~1.05 KB/1076, heapSize ~1.36 KB/1392, currentSize=0 B/0 for 26be18c713c8da53cab1008ab6b5dbbc in 28ms, sequenceid=22, compaction requested=true 2024-12-05T15:16:39,630 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/3e70d4efe9924928826be8a9b845500a, hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/c725144f29074ffab03d5aa502ec88d8, hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/bf704d6b87f9409b852e4f0bda67f40a] to archive 2024-12-05T15:16:39,631 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-05T15:16:39,632 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/3e70d4efe9924928826be8a9b845500a to hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/3e70d4efe9924928826be8a9b845500a 2024-12-05T15:16:39,634 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/c725144f29074ffab03d5aa502ec88d8 to hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/c725144f29074ffab03d5aa502ec88d8 2024-12-05T15:16:39,635 DEBUG [StoreCloser-TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/bf704d6b87f9409b852e4f0bda67f40a to hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/archive/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/info/bf704d6b87f9409b852e4f0bda67f40a 2024-12-05T15:16:39,639 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/default/TestLogRolling-testCompactionRecordDoesntBlockRolling/26be18c713c8da53cab1008ab6b5dbbc/recovered.edits/25.seqid, newMaxSeqId=25, maxSeqId=1 2024-12-05T15:16:39,640 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:39,640 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 26be18c713c8da53cab1008ab6b5dbbc: 2024-12-05T15:16:39,640 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testCompactionRecordDoesntBlockRolling,,1733411739443.26be18c713c8da53cab1008ab6b5dbbc. 2024-12-05T15:16:39,646 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/.tmp/table/bb65fbaa6ae540d7b3bfcd5279e98f71 is 89, key is TestLogRolling-testCompactionRecordDoesntBlockRolling/table:state/1733411739802/Put/seqid=0 2024-12-05T15:16:39,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741850_1026 (size=5532) 2024-12-05T15:16:39,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741850_1026 (size=5532) 2024-12-05T15:16:39,651 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=264 B at sequenceid=14 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/.tmp/table/bb65fbaa6ae540d7b3bfcd5279e98f71 2024-12-05T15:16:39,656 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/.tmp/info/cc43432509cf46e8b42abdfd237c3a82 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/info/cc43432509cf46e8b42abdfd237c3a82 2024-12-05T15:16:39,661 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/info/cc43432509cf46e8b42abdfd237c3a82, entries=20, sequenceid=14, filesize=8.2 K 2024-12-05T15:16:39,662 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/.tmp/table/bb65fbaa6ae540d7b3bfcd5279e98f71 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/table/bb65fbaa6ae540d7b3bfcd5279e98f71 2024-12-05T15:16:39,666 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/table/bb65fbaa6ae540d7b3bfcd5279e98f71, entries=4, sequenceid=14, filesize=5.4 K 2024-12-05T15:16:39,667 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~3.05 KB/3122, heapSize ~5.27 KB/5400, currentSize=0 B/0 for 1588230740 in 70ms, sequenceid=14, compaction requested=false 2024-12-05T15:16:39,671 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/data/hbase/meta/1588230740/recovered.edits/17.seqid, newMaxSeqId=17, maxSeqId=1 2024-12-05T15:16:39,672 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T15:16:39,672 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:16:39,672 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:16:39,672 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-05T15:16:39,797 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,33595,1733411738098; all regions closed. 2024-12-05T15:16:39,797 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098 2024-12-05T15:16:39,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741834_1010 (size=4570) 2024-12-05T15:16:39,799 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741834_1010 (size=4570) 2024-12-05T15:16:39,802 DEBUG [RS:0;4ecb0f57b706:33595 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/oldWALs 2024-12-05T15:16:39,802 INFO [RS:0;4ecb0f57b706:33595 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C33595%2C1733411738098.meta:.meta(num 1733411738887) 2024-12-05T15:16:39,802 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/WALs/4ecb0f57b706,33595,1733411738098 2024-12-05T15:16:39,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741847_1023 (size=1545) 2024-12-05T15:16:39,804 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741847_1023 (size=1545) 2024-12-05T15:16:39,806 DEBUG [RS:0;4ecb0f57b706:33595 {}] wal.AbstractFSWAL(1071): Moved 2 WAL file(s) to /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/oldWALs 2024-12-05T15:16:39,806 INFO [RS:0;4ecb0f57b706:33595 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C33595%2C1733411738098:(num 1733411799585) 2024-12-05T15:16:39,806 DEBUG [RS:0;4ecb0f57b706:33595 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:16:39,806 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:16:39,807 INFO [RS:0;4ecb0f57b706:33595 {}] hbase.ChoreService(370): Chore service for: regionserver/4ecb0f57b706:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-05T15:16:39,807 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:16:39,807 INFO [RS:0;4ecb0f57b706:33595 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:33595 2024-12-05T15:16:39,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:16:39,809 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/4ecb0f57b706,33595,1733411738098 2024-12-05T15:16:39,810 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [4ecb0f57b706,33595,1733411738098] 2024-12-05T15:16:39,810 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 4ecb0f57b706,33595,1733411738098; numProcessing=1 2024-12-05T15:16:39,811 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/4ecb0f57b706,33595,1733411738098 already deleted, retry=false 2024-12-05T15:16:39,811 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 4ecb0f57b706,33595,1733411738098 expired; onlineServers=0 2024-12-05T15:16:39,811 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,39961,1733411738024' ***** 2024-12-05T15:16:39,811 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T15:16:39,811 DEBUG [M:0;4ecb0f57b706:39961 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@71cab50b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:16:39,811 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,39961,1733411738024 2024-12-05T15:16:39,812 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,39961,1733411738024; all regions closed. 2024-12-05T15:16:39,812 DEBUG [M:0;4ecb0f57b706:39961 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:16:39,812 DEBUG [M:0;4ecb0f57b706:39961 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T15:16:39,812 DEBUG [M:0;4ecb0f57b706:39961 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T15:16:39,812 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T15:16:39,812 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411738275 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411738275,5,FailOnTimeoutGroup] 2024-12-05T15:16:39,812 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411738271 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411738271,5,FailOnTimeoutGroup] 2024-12-05T15:16:39,812 INFO [M:0;4ecb0f57b706:39961 {}] hbase.ChoreService(370): Chore service for: master/4ecb0f57b706:0 had [] on shutdown 2024-12-05T15:16:39,812 DEBUG [M:0;4ecb0f57b706:39961 {}] master.HMaster(1733): Stopping service threads 2024-12-05T15:16:39,812 INFO [M:0;4ecb0f57b706:39961 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T15:16:39,813 INFO [M:0;4ecb0f57b706:39961 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T15:16:39,813 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T15:16:39,813 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T15:16:39,813 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:39,813 DEBUG [M:0;4ecb0f57b706:39961 {}] zookeeper.ZKUtil(347): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T15:16:39,813 WARN [M:0;4ecb0f57b706:39961 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T15:16:39,813 INFO [M:0;4ecb0f57b706:39961 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-05T15:16:39,813 INFO [M:0;4ecb0f57b706:39961 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T15:16:39,813 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:16:39,813 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:39,813 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:16:39,814 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:39,814 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:16:39,814 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:39,814 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=65.09 KB heapSize=81.71 KB 2024-12-05T15:16:39,831 DEBUG [M:0;4ecb0f57b706:39961 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f0d5e8efd4ee459f8de2172a7a2f0b4a is 82, key is hbase:meta,,1/info:regioninfo/1733411738906/Put/seqid=0 2024-12-05T15:16:39,836 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741851_1027 (size=5672) 2024-12-05T15:16:39,836 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741851_1027 (size=5672) 2024-12-05T15:16:39,837 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f0d5e8efd4ee459f8de2172a7a2f0b4a 2024-12-05T15:16:39,856 DEBUG [M:0;4ecb0f57b706:39961 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/76ba795a65274977b31cd47d5ed66825 is 799, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733411739808/Put/seqid=0 2024-12-05T15:16:39,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741852_1028 (size=8357) 2024-12-05T15:16:39,862 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741852_1028 (size=8357) 2024-12-05T15:16:39,862 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=64.48 KB at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/76ba795a65274977b31cd47d5ed66825 2024-12-05T15:16:39,867 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 76ba795a65274977b31cd47d5ed66825 2024-12-05T15:16:39,881 DEBUG [M:0;4ecb0f57b706:39961 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6ccdf43dfdd94dcd8706078c7c2e2563 is 69, key is 4ecb0f57b706,33595,1733411738098/rs:state/1733411738354/Put/seqid=0 2024-12-05T15:16:39,886 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741853_1029 (size=5156) 2024-12-05T15:16:39,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741853_1029 (size=5156) 2024-12-05T15:16:39,887 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6ccdf43dfdd94dcd8706078c7c2e2563 2024-12-05T15:16:39,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:16:39,911 DEBUG [M:0;4ecb0f57b706:39961 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/36f25a3cd2054b76a91d5c6ade76ef41 is 52, key is load_balancer_on/state:d/1733411739437/Put/seqid=0 2024-12-05T15:16:39,911 INFO [RS:0;4ecb0f57b706:33595 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,33595,1733411738098; zookeeper connection closed. 2024-12-05T15:16:39,911 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:33595-0x10062bfca7b0001, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:16:39,911 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@257c4c77 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@257c4c77 2024-12-05T15:16:39,911 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-05T15:16:39,916 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741854_1030 (size=5056) 2024-12-05T15:16:39,916 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741854_1030 (size=5056) 2024-12-05T15:16:39,917 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=184 (bloomFilter=true), to=hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/36f25a3cd2054b76a91d5c6ade76ef41 2024-12-05T15:16:39,922 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/f0d5e8efd4ee459f8de2172a7a2f0b4a as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/f0d5e8efd4ee459f8de2172a7a2f0b4a 2024-12-05T15:16:39,927 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/f0d5e8efd4ee459f8de2172a7a2f0b4a, entries=8, sequenceid=184, filesize=5.5 K 2024-12-05T15:16:39,928 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/76ba795a65274977b31cd47d5ed66825 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/76ba795a65274977b31cd47d5ed66825 2024-12-05T15:16:39,933 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for 76ba795a65274977b31cd47d5ed66825 2024-12-05T15:16:39,933 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/76ba795a65274977b31cd47d5ed66825, entries=21, sequenceid=184, filesize=8.2 K 2024-12-05T15:16:39,934 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/6ccdf43dfdd94dcd8706078c7c2e2563 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/6ccdf43dfdd94dcd8706078c7c2e2563 2024-12-05T15:16:39,939 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/6ccdf43dfdd94dcd8706078c7c2e2563, entries=1, sequenceid=184, filesize=5.0 K 2024-12-05T15:16:39,940 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/36f25a3cd2054b76a91d5c6ade76ef41 as hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/36f25a3cd2054b76a91d5c6ade76ef41 2024-12-05T15:16:39,945 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:38195/user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/36f25a3cd2054b76a91d5c6ade76ef41, entries=1, sequenceid=184, filesize=4.9 K 2024-12-05T15:16:39,946 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(3040): Finished flush of dataSize ~65.09 KB/66649, heapSize ~81.65 KB/83608, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 132ms, sequenceid=184, compaction requested=false 2024-12-05T15:16:39,947 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:39,947 DEBUG [M:0;4ecb0f57b706:39961 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:16:39,948 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/621b1e4b-af5f-24dd-d0d8-4a66050e45c4/MasterData/WALs/4ecb0f57b706,39961,1733411738024 2024-12-05T15:16:39,949 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:37029 is added to blk_1073741830_1006 (size=79170) 2024-12-05T15:16:39,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:45861 is added to blk_1073741830_1006 (size=79170) 2024-12-05T15:16:39,950 INFO [M:0;4ecb0f57b706:39961 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-05T15:16:39,950 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:16:39,950 INFO [M:0;4ecb0f57b706:39961 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:39961 2024-12-05T15:16:39,952 DEBUG [M:0;4ecb0f57b706:39961 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/4ecb0f57b706,39961,1733411738024 already deleted, retry=false 2024-12-05T15:16:40,053 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:16:40,053 INFO [M:0;4ecb0f57b706:39961 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,39961,1733411738024; zookeeper connection closed. 2024-12-05T15:16:40,053 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39961-0x10062bfca7b0000, quorum=127.0.0.1:63920, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:16:40,061 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@55215408{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:16:40,061 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@262a1159{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:16:40,061 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:16:40,061 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5d858d58{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:16:40,061 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3ac49840{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir/,STOPPED} 2024-12-05T15:16:40,063 WARN [BP-1679678351-172.17.0.2-1733411737314 heartbeating to localhost/127.0.0.1:38195 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:16:40,063 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:16:40,063 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:16:40,063 WARN [BP-1679678351-172.17.0.2-1733411737314 heartbeating to localhost/127.0.0.1:38195 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1679678351-172.17.0.2-1733411737314 (Datanode Uuid 2a2cfa40-e6f8-4a02-98fb-1fc304ca18b9) service to localhost/127.0.0.1:38195 2024-12-05T15:16:40,064 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data3/current/BP-1679678351-172.17.0.2-1733411737314 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:16:40,064 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data4/current/BP-1679678351-172.17.0.2-1733411737314 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:16:40,064 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:16:40,066 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4d38e2e3{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:16:40,066 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@5728e0aa{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:16:40,066 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:16:40,067 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@2899e0b7{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:16:40,067 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@79eaf123{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir/,STOPPED} 2024-12-05T15:16:40,068 WARN [BP-1679678351-172.17.0.2-1733411737314 heartbeating to localhost/127.0.0.1:38195 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:16:40,068 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:16:40,068 WARN [BP-1679678351-172.17.0.2-1733411737314 heartbeating to localhost/127.0.0.1:38195 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1679678351-172.17.0.2-1733411737314 (Datanode Uuid 06c36c51-7269-48cf-9467-b9873b0aa699) service to localhost/127.0.0.1:38195 2024-12-05T15:16:40,068 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:16:40,069 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data1/current/BP-1679678351-172.17.0.2-1733411737314 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:16:40,069 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/cluster_920e5170-cd01-6340-63d8-e7f317b358e6/dfs/data/data2/current/BP-1679678351-172.17.0.2-1733411737314 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:16:40,069 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:16:40,075 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@307f11b9{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:16:40,076 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6bcef121{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:16:40,076 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:16:40,076 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@5559bf37{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:16:40,076 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7cea72ca{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir/,STOPPED} 2024-12-05T15:16:40,082 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-05T15:16:40,102 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-05T15:16:40,110 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testCompactionRecordDoesntBlockRolling Thread=112 (was 101) - Thread LEAK? -, OpenFileDescriptor=467 (was 444) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=130 (was 130), ProcessCount=11 (was 11), AvailableMemoryMB=8425 (was 8568) 2024-12-05T15:16:40,117 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRolling Thread=113, OpenFileDescriptor=467, MaxFileDescriptor=1048576, SystemLoadAverage=130, ProcessCount=11, AvailableMemoryMB=8425 2024-12-05T15:16:40,117 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T15:16:40,117 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.log.dir so I do NOT create it in target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5 2024-12-05T15:16:40,117 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/dd1d0ffa-b0ee-ba51-0778-6ad2ba3ec693/hadoop.tmp.dir so I do NOT create it in target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5 2024-12-05T15:16:40,117 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083, deleteOnExit=true 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/test.cache.data in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-05T15:16:40,118 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:16:40,118 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/nfs.dump.dir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/java.io.tmpdir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T15:16:40,119 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T15:16:40,132 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:16:40,167 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:40,190 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:16:40,194 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:16:40,195 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:16:40,195 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:16:40,195 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:16:40,195 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:16:40,196 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3378fb3{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:16:40,196 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@32cfb9e6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:16:40,310 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@17dd071d{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/java.io.tmpdir/jetty-localhost-40871-hadoop-hdfs-3_4_1-tests_jar-_-any-1198286007886138818/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:16:40,311 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@23ed981c{HTTP/1.1, (http/1.1)}{localhost:40871} 2024-12-05T15:16:40,311 INFO [Time-limited test {}] server.Server(415): Started @291126ms 2024-12-05T15:16:40,324 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:16:40,342 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:40,372 INFO [regionserver/4ecb0f57b706:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:16:40,377 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:16:40,380 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:16:40,380 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:16:40,380 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:16:40,380 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2024-12-05T15:16:40,381 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@15df8ed8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:16:40,381 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@50bb1e61{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:16:40,495 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@555a9157{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/java.io.tmpdir/jetty-localhost-36887-hadoop-hdfs-3_4_1-tests_jar-_-any-13147849678157102039/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:16:40,496 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@86b8092{HTTP/1.1, (http/1.1)}{localhost:36887} 2024-12-05T15:16:40,496 INFO [Time-limited test {}] server.Server(415): Started @291311ms 2024-12-05T15:16:40,497 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:16:40,526 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:16:40,529 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:16:40,529 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:16:40,529 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:16:40,529 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:16:40,530 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@325f9d5d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:16:40,530 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@65cd88a1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:16:40,593 WARN [Thread-1696 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data2/current/BP-1838888116-172.17.0.2-1733411800140/current, will proceed with Du for space computation calculation, 2024-12-05T15:16:40,593 WARN [Thread-1695 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data1/current/BP-1838888116-172.17.0.2-1733411800140/current, will proceed with Du for space computation calculation, 2024-12-05T15:16:40,617 WARN [Thread-1674 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:16:40,620 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7a6a9b5f989501e6 with lease ID 0xf8f3d68a90af1c2: Processing first storage report for DS-1ac9a4cb-2038-429e-a43b-71bf04cba9a4 from datanode DatanodeRegistration(127.0.0.1:43711, datanodeUuid=903ba1ae-98e8-488f-806e-3e814922ab8b, infoPort=41935, infoSecurePort=0, ipcPort=40723, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140) 2024-12-05T15:16:40,620 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7a6a9b5f989501e6 with lease ID 0xf8f3d68a90af1c2: from storage DS-1ac9a4cb-2038-429e-a43b-71bf04cba9a4 node DatanodeRegistration(127.0.0.1:43711, datanodeUuid=903ba1ae-98e8-488f-806e-3e814922ab8b, infoPort=41935, infoSecurePort=0, ipcPort=40723, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:16:40,620 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7a6a9b5f989501e6 with lease ID 0xf8f3d68a90af1c2: Processing first storage report for DS-be341d2c-ffa0-48cb-8b75-b44c46f75592 from datanode DatanodeRegistration(127.0.0.1:43711, datanodeUuid=903ba1ae-98e8-488f-806e-3e814922ab8b, infoPort=41935, infoSecurePort=0, ipcPort=40723, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140) 2024-12-05T15:16:40,620 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7a6a9b5f989501e6 with lease ID 0xf8f3d68a90af1c2: from storage DS-be341d2c-ffa0-48cb-8b75-b44c46f75592 node DatanodeRegistration(127.0.0.1:43711, datanodeUuid=903ba1ae-98e8-488f-806e-3e814922ab8b, infoPort=41935, infoSecurePort=0, ipcPort=40723, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:16:40,652 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@33dd8821{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/java.io.tmpdir/jetty-localhost-34827-hadoop-hdfs-3_4_1-tests_jar-_-any-3574323792202601704/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:16:40,652 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@6e41ff6b{HTTP/1.1, (http/1.1)}{localhost:34827} 2024-12-05T15:16:40,653 INFO [Time-limited test {}] server.Server(415): Started @291468ms 2024-12-05T15:16:40,654 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:16:40,746 WARN [Thread-1721 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data3/current/BP-1838888116-172.17.0.2-1733411800140/current, will proceed with Du for space computation calculation, 2024-12-05T15:16:40,746 WARN [Thread-1722 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data4/current/BP-1838888116-172.17.0.2-1733411800140/current, will proceed with Du for space computation calculation, 2024-12-05T15:16:40,765 WARN [Thread-1710 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:16:40,767 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7e1a6019d4636fa1 with lease ID 0xf8f3d68a90af1c3: Processing first storage report for DS-3a2bea15-0dbf-4668-b025-48eddcf2bb10 from datanode DatanodeRegistration(127.0.0.1:39039, datanodeUuid=59a69e00-1e13-44f8-81d6-b064a4e379ba, infoPort=39577, infoSecurePort=0, ipcPort=46689, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140) 2024-12-05T15:16:40,767 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7e1a6019d4636fa1 with lease ID 0xf8f3d68a90af1c3: from storage DS-3a2bea15-0dbf-4668-b025-48eddcf2bb10 node DatanodeRegistration(127.0.0.1:39039, datanodeUuid=59a69e00-1e13-44f8-81d6-b064a4e379ba, infoPort=39577, infoSecurePort=0, ipcPort=46689, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2024-12-05T15:16:40,767 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7e1a6019d4636fa1 with lease ID 0xf8f3d68a90af1c3: Processing first storage report for DS-a860289b-0bbd-4b7c-8271-47a293992c1d from datanode DatanodeRegistration(127.0.0.1:39039, datanodeUuid=59a69e00-1e13-44f8-81d6-b064a4e379ba, infoPort=39577, infoSecurePort=0, ipcPort=46689, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140) 2024-12-05T15:16:40,767 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7e1a6019d4636fa1 with lease ID 0xf8f3d68a90af1c3: from storage DS-a860289b-0bbd-4b7c-8271-47a293992c1d node DatanodeRegistration(127.0.0.1:39039, datanodeUuid=59a69e00-1e13-44f8-81d6-b064a4e379ba, infoPort=39577, infoSecurePort=0, ipcPort=46689, storageInfo=lv=-57;cid=testClusterID;nsid=902818193;c=1733411800140), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:16:40,777 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5 2024-12-05T15:16:40,779 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/zookeeper_0, clientPort=59287, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T15:16:40,780 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=59287 2024-12-05T15:16:40,780 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:40,781 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:40,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:16:40,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:16:40,790 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2 with version=8 2024-12-05T15:16:40,790 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/hbase-staging 2024-12-05T15:16:40,792 INFO [Time-limited test {}] client.ConnectionUtils(129): master/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:16:40,792 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:16:40,792 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:16:40,792 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:16:40,792 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:16:40,792 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:16:40,792 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:16:40,792 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:16:40,793 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:42659 2024-12-05T15:16:40,793 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:40,794 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:40,796 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:42659 connecting to ZooKeeper ensemble=127.0.0.1:59287 2024-12-05T15:16:40,802 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:426590x0, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:16:40,802 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:42659-0x10062c0bfb10000 connected 2024-12-05T15:16:40,816 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:16:40,816 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:16:40,817 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:16:40,817 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=42659 2024-12-05T15:16:40,817 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=42659 2024-12-05T15:16:40,817 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=42659 2024-12-05T15:16:40,818 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=42659 2024-12-05T15:16:40,818 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=42659 2024-12-05T15:16:40,818 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2, hbase.cluster.distributed=false 2024-12-05T15:16:40,834 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:16:40,834 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:16:40,834 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:16:40,834 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:16:40,834 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:16:40,834 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:16:40,834 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:16:40,835 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:16:40,835 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:43509 2024-12-05T15:16:40,835 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T15:16:40,836 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T15:16:40,836 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:40,838 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:40,840 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:43509 connecting to ZooKeeper ensemble=127.0.0.1:59287 2024-12-05T15:16:40,843 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:435090x0, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:16:40,843 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:435090x0, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:16:40,843 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:43509-0x10062c0bfb10001 connected 2024-12-05T15:16:40,843 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:16:40,844 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:16:40,844 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=43509 2024-12-05T15:16:40,844 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=43509 2024-12-05T15:16:40,847 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=43509 2024-12-05T15:16:40,848 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=43509 2024-12-05T15:16:40,848 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=43509 2024-12-05T15:16:40,849 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:40,850 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:16:40,850 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:16:40,851 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:40,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:16:40,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:16:40,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:40,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:40,854 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:16:40,854 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/4ecb0f57b706,42659,1733411800791 from backup master directory 2024-12-05T15:16:40,854 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:16:40,856 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:40,856 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:16:40,856 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:16:40,856 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:16:40,856 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:40,861 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;4ecb0f57b706:42659 2024-12-05T15:16:40,869 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:16:40,870 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:16:40,871 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/hbase.id with ID: 71c1257f-2fe6-4bb3-8c8b-a7c212fc7d39 2024-12-05T15:16:40,880 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:40,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:40,883 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:40,893 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:16:40,893 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:16:40,894 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:16:40,895 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T15:16:40,895 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:16:40,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:16:40,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:16:40,902 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store 2024-12-05T15:16:40,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:16:40,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:16:41,168 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:41,310 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:41,310 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:16:41,310 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:41,310 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:41,310 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:16:41,310 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:41,310 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:16:41,310 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:16:41,311 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/.initializing 2024-12-05T15:16:41,311 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/WALs/4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:41,314 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C42659%2C1733411800791, suffix=, logDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/WALs/4ecb0f57b706,42659,1733411800791, archiveDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/oldWALs, maxLogs=10 2024-12-05T15:16:41,314 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C42659%2C1733411800791.1733411801314 2024-12-05T15:16:41,319 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/WALs/4ecb0f57b706,42659,1733411800791/4ecb0f57b706%2C42659%2C1733411800791.1733411801314 2024-12-05T15:16:41,319 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:39577:39577),(127.0.0.1/127.0.0.1:41935:41935)] 2024-12-05T15:16:41,320 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:16:41,320 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:41,320 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,320 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,321 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,322 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T15:16:41,322 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,323 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:16:41,323 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,324 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T15:16:41,324 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,324 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:16:41,324 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,325 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T15:16:41,325 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,326 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:16:41,326 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,327 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T15:16:41,327 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,327 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:16:41,328 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,328 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,330 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T15:16:41,331 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:16:41,333 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:16:41,333 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=753978, jitterRate=-0.041267409920692444}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T15:16:41,334 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:16:41,334 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T15:16:41,338 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@503373e7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:16:41,338 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-05T15:16:41,339 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T15:16:41,339 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T15:16:41,339 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T15:16:41,339 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-05T15:16:41,339 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-05T15:16:41,340 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T15:16:41,341 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T15:16:41,342 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T15:16:41,343 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:41,344 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-05T15:16:41,344 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T15:16:41,344 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T15:16:41,346 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-05T15:16:41,346 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T15:16:41,347 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T15:16:41,348 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-05T15:16:41,349 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T15:16:41,350 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T15:16:41,351 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T15:16:41,353 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T15:16:41,354 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:16:41,354 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:16:41,354 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,354 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,355 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=4ecb0f57b706,42659,1733411800791, sessionid=0x10062c0bfb10000, setting cluster-up flag (Was=false) 2024-12-05T15:16:41,358 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,358 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,363 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T15:16:41,364 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:41,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,367 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,372 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T15:16:41,373 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:41,375 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-05T15:16:41,375 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-05T15:16:41,376 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 4ecb0f57b706,42659,1733411800791 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/4ecb0f57b706:0, corePoolSize=10, maxPoolSize=10 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:16:41,376 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,381 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:16:41,381 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-05T15:16:41,381 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733411831381 2024-12-05T15:16:41,381 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T15:16:41,381 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T15:16:41,381 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T15:16:41,381 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T15:16:41,381 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T15:16:41,381 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T15:16:41,382 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,382 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T15:16:41,382 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T15:16:41,382 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T15:16:41,382 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,382 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T15:16:41,382 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T15:16:41,382 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:16:41,382 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411801382,5,FailOnTimeoutGroup] 2024-12-05T15:16:41,383 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411801382,5,FailOnTimeoutGroup] 2024-12-05T15:16:41,383 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,383 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T15:16:41,383 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,383 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:16:41,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:16:41,391 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-05T15:16:41,392 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2 2024-12-05T15:16:41,397 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:16:41,398 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:16:41,398 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:41,399 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:16:41,400 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:16:41,400 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,401 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:16:41,401 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:16:41,402 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:16:41,402 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,402 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:16:41,403 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:16:41,403 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:16:41,404 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,404 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:16:41,405 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740 2024-12-05T15:16:41,405 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740 2024-12-05T15:16:41,406 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:16:41,407 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:16:41,409 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:16:41,409 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=736194, jitterRate=-0.0638810396194458}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:16:41,409 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:16:41,409 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:16:41,409 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:16:41,409 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:16:41,409 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:16:41,409 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:16:41,410 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:16:41,410 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:16:41,411 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:16:41,411 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-05T15:16:41,411 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T15:16:41,412 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T15:16:41,413 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T15:16:41,461 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;4ecb0f57b706:43509 2024-12-05T15:16:41,462 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1008): ClusterId : 71c1257f-2fe6-4bb3-8c8b-a7c212fc7d39 2024-12-05T15:16:41,462 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T15:16:41,464 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T15:16:41,464 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T15:16:41,466 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T15:16:41,466 DEBUG [RS:0;4ecb0f57b706:43509 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@8113f5f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:16:41,467 DEBUG [RS:0;4ecb0f57b706:43509 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@74925ace, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:16:41,467 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-05T15:16:41,467 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-05T15:16:41,467 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-05T15:16:41,467 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,42659,1733411800791 with isa=4ecb0f57b706/172.17.0.2:43509, startcode=1733411800834 2024-12-05T15:16:41,468 DEBUG [RS:0;4ecb0f57b706:43509 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:16:41,470 INFO [RS-EventLoopGroup-12-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:45471, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:16:41,470 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42659 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,470 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42659 {}] master.ServerManager(486): Registering regionserver=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,471 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2 2024-12-05T15:16:41,471 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:34657 2024-12-05T15:16:41,472 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-05T15:16:41,474 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:16:41,474 DEBUG [RS:0;4ecb0f57b706:43509 {}] zookeeper.ZKUtil(111): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,474 WARN [RS:0;4ecb0f57b706:43509 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:16:41,474 INFO [RS:0;4ecb0f57b706:43509 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:16:41,474 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,475 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [4ecb0f57b706,43509,1733411800834] 2024-12-05T15:16:41,482 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-05T15:16:41,483 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T15:16:41,487 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T15:16:41,488 INFO [RS:0;4ecb0f57b706:43509 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:16:41,488 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-05T15:16:41,489 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:16:41,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:16:41,490 DEBUG [RS:0;4ecb0f57b706:43509 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:16:41,491 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,491 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,491 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,491 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,491 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,43509,1733411800834-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:16:41,515 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T15:16:41,515 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,43509,1733411800834-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:41,534 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.Replication(204): 4ecb0f57b706,43509,1733411800834 started 2024-12-05T15:16:41,535 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1767): Serving as 4ecb0f57b706,43509,1733411800834, RpcServer on 4ecb0f57b706/172.17.0.2:43509, sessionid=0x10062c0bfb10001 2024-12-05T15:16:41,535 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T15:16:41,535 DEBUG [RS:0;4ecb0f57b706:43509 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,535 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,43509,1733411800834' 2024-12-05T15:16:41,535 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T15:16:41,535 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T15:16:41,536 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T15:16:41,536 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T15:16:41,536 DEBUG [RS:0;4ecb0f57b706:43509 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,536 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,43509,1733411800834' 2024-12-05T15:16:41,536 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T15:16:41,536 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T15:16:41,536 DEBUG [RS:0;4ecb0f57b706:43509 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T15:16:41,536 INFO [RS:0;4ecb0f57b706:43509 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T15:16:41,536 INFO [RS:0;4ecb0f57b706:43509 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T15:16:41,563 WARN [4ecb0f57b706:42659 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-05T15:16:41,638 INFO [RS:0;4ecb0f57b706:43509 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C43509%2C1733411800834, suffix=, logDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834, archiveDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/oldWALs, maxLogs=32 2024-12-05T15:16:41,638 INFO [RS:0;4ecb0f57b706:43509 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C43509%2C1733411800834.1733411801638 2024-12-05T15:16:41,644 INFO [RS:0;4ecb0f57b706:43509 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411801638 2024-12-05T15:16:41,644 DEBUG [RS:0;4ecb0f57b706:43509 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41935:41935),(127.0.0.1/127.0.0.1:39577:39577)] 2024-12-05T15:16:41,813 DEBUG [4ecb0f57b706:42659 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-05T15:16:41,813 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,814 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,43509,1733411800834, state=OPENING 2024-12-05T15:16:41,816 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T15:16:41,818 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,818 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:41,818 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=4ecb0f57b706,43509,1733411800834}] 2024-12-05T15:16:41,818 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:16:41,818 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:16:41,971 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,971 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T15:16:41,973 INFO [RS-EventLoopGroup-13-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56152, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T15:16:41,976 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-05T15:16:41,976 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:16:41,978 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C43509%2C1733411800834.meta, suffix=.meta, logDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834, archiveDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/oldWALs, maxLogs=32 2024-12-05T15:16:41,978 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C43509%2C1733411800834.meta.1733411801978.meta 2024-12-05T15:16:41,983 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.meta.1733411801978.meta 2024-12-05T15:16:41,983 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41935:41935),(127.0.0.1/127.0.0.1:39577:39577)] 2024-12-05T15:16:41,983 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:16:41,984 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T15:16:41,984 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T15:16:41,984 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T15:16:41,984 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T15:16:41,984 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:41,984 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-05T15:16:41,984 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-05T15:16:41,985 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:16:41,986 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:16:41,986 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,986 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:16:41,987 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:16:41,987 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:16:41,987 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,988 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:16:41,988 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:16:41,988 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:16:41,988 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:41,989 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:16:41,989 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740 2024-12-05T15:16:41,990 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740 2024-12-05T15:16:41,991 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:16:41,992 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:16:41,993 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=743292, jitterRate=-0.054856061935424805}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:16:41,993 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:16:41,994 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733411801971 2024-12-05T15:16:41,995 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T15:16:41,995 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-05T15:16:41,996 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:41,997 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,43509,1733411800834, state=OPEN 2024-12-05T15:16:42,001 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:16:42,001 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:16:42,001 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:16:42,002 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:16:42,003 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T15:16:42,003 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=4ecb0f57b706,43509,1733411800834 in 183 msec 2024-12-05T15:16:42,005 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T15:16:42,005 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 592 msec 2024-12-05T15:16:42,006 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 631 msec 2024-12-05T15:16:42,007 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733411802006, completionTime=-1 2024-12-05T15:16:42,007 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-05T15:16:42,007 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-05T15:16:42,007 DEBUG [hconnection-0x2e0a4258-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:16:42,009 INFO [RS-EventLoopGroup-13-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56156, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:16:42,010 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-05T15:16:42,010 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733411862010 2024-12-05T15:16:42,010 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733411922010 2024-12-05T15:16:42,010 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 3 msec 2024-12-05T15:16:42,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,42659,1733411800791-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:42,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,42659,1733411800791-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:42,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,42659,1733411800791-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:42,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-4ecb0f57b706:42659, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:42,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:42,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-05T15:16:42,016 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:16:42,017 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-05T15:16:42,017 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-05T15:16:42,018 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:16:42,018 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:42,019 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:16:42,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:16:42,025 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:16:42,026 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 14ad900f35513e0519981bcd843ee526, NAME => 'hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2 2024-12-05T15:16:42,031 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:16:42,032 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:16:42,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:42,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 14ad900f35513e0519981bcd843ee526, disabling compactions & flushes 2024-12-05T15:16:42,032 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:16:42,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:16:42,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. after waiting 0 ms 2024-12-05T15:16:42,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:16:42,032 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:16:42,032 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 14ad900f35513e0519981bcd843ee526: 2024-12-05T15:16:42,033 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:16:42,034 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733411802033"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411802033"}]},"ts":"1733411802033"} 2024-12-05T15:16:42,035 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:16:42,036 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:16:42,036 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411802036"}]},"ts":"1733411802036"} 2024-12-05T15:16:42,037 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-05T15:16:42,042 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=14ad900f35513e0519981bcd843ee526, ASSIGN}] 2024-12-05T15:16:42,043 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=14ad900f35513e0519981bcd843ee526, ASSIGN 2024-12-05T15:16:42,043 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=14ad900f35513e0519981bcd843ee526, ASSIGN; state=OFFLINE, location=4ecb0f57b706,43509,1733411800834; forceNewPlan=false, retain=false 2024-12-05T15:16:42,168 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:42,194 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=14ad900f35513e0519981bcd843ee526, regionState=OPENING, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:42,196 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 14ad900f35513e0519981bcd843ee526, server=4ecb0f57b706,43509,1733411800834}] 2024-12-05T15:16:42,343 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:42,348 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:42,352 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:16:42,352 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 14ad900f35513e0519981bcd843ee526, NAME => 'hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:16:42,352 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 14ad900f35513e0519981bcd843ee526 2024-12-05T15:16:42,352 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:42,352 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 14ad900f35513e0519981bcd843ee526 2024-12-05T15:16:42,352 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 14ad900f35513e0519981bcd843ee526 2024-12-05T15:16:42,354 INFO [StoreOpener-14ad900f35513e0519981bcd843ee526-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 14ad900f35513e0519981bcd843ee526 2024-12-05T15:16:42,355 INFO [StoreOpener-14ad900f35513e0519981bcd843ee526-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 14ad900f35513e0519981bcd843ee526 columnFamilyName info 2024-12-05T15:16:42,355 DEBUG [StoreOpener-14ad900f35513e0519981bcd843ee526-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:42,355 INFO [StoreOpener-14ad900f35513e0519981bcd843ee526-1 {}] regionserver.HStore(327): Store=14ad900f35513e0519981bcd843ee526/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:16:42,356 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526 2024-12-05T15:16:42,356 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526 2024-12-05T15:16:42,358 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 14ad900f35513e0519981bcd843ee526 2024-12-05T15:16:42,360 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:16:42,361 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 14ad900f35513e0519981bcd843ee526; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=753936, jitterRate=-0.041321784257888794}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:16:42,361 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 14ad900f35513e0519981bcd843ee526: 2024-12-05T15:16:42,362 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526., pid=6, masterSystemTime=1733411802348 2024-12-05T15:16:42,363 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:16:42,363 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:16:42,364 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=14ad900f35513e0519981bcd843ee526, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:42,367 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T15:16:42,367 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 14ad900f35513e0519981bcd843ee526, server=4ecb0f57b706,43509,1733411800834 in 169 msec 2024-12-05T15:16:42,369 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T15:16:42,369 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=14ad900f35513e0519981bcd843ee526, ASSIGN in 325 msec 2024-12-05T15:16:42,369 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:16:42,370 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411802369"}]},"ts":"1733411802369"} 2024-12-05T15:16:42,371 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-05T15:16:42,373 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:16:42,375 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 357 msec 2024-12-05T15:16:42,418 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-05T15:16:42,420 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:42,420 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:16:42,420 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:16:42,424 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-05T15:16:42,432 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:16:42,435 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 11 msec 2024-12-05T15:16:42,446 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-05T15:16:42,453 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:16:42,456 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 10 msec 2024-12-05T15:16:42,461 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-05T15:16:42,463 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-05T15:16:42,463 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.607sec 2024-12-05T15:16:42,463 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T15:16:42,463 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T15:16:42,463 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T15:16:42,464 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T15:16:42,464 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T15:16:42,464 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,42659,1733411800791-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:16:42,464 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,42659,1733411800791-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T15:16:42,465 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-05T15:16:42,465 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T15:16:42,465 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,42659,1733411800791-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:16:42,551 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7319ca6d to 127.0.0.1:59287 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@29df1205 2024-12-05T15:16:42,555 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7ec62145, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:16:42,557 DEBUG [hconnection-0x1c801215-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:16:42,559 INFO [RS-EventLoopGroup-13-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:56170, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:16:42,560 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=4ecb0f57b706,42659,1733411800791 2024-12-05T15:16:42,560 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:16:42,562 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-05T15:16:42,563 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2024-12-05T15:16:42,565 INFO [RS-EventLoopGroup-12-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:48322, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2024-12-05T15:16:42,566 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42659 {}] util.TableDescriptorChecker(321): MAX_FILESIZE for table descriptor or "hbase.hregion.max.filesize" (786432) is too small, which might cause over splitting into unmanageable number of regions. 2024-12-05T15:16:42,566 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42659 {}] util.TableDescriptorChecker(321): MEMSTORE_FLUSHSIZE for table descriptor or "hbase.hregion.memstore.flush.size" (8192) is too small, which might cause very frequent flushing. 2024-12-05T15:16:42,566 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42659 {}] master.HMaster$4(2389): Client=jenkins//172.17.0.2 create 'TestLogRolling-testLogRolling', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:16:42,567 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42659 {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=TestLogRolling-testLogRolling 2024-12-05T15:16:42,568 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:16:42,568 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:42,568 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42659 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.2 procedure request for creating table: namespace: "default" qualifier: "TestLogRolling-testLogRolling" procId is: 9 2024-12-05T15:16:42,569 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:16:42,569 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42659 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:16:42,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741837_1013 (size=381) 2024-12-05T15:16:42,576 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741837_1013 (size=381) 2024-12-05T15:16:42,578 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 7321143ee598c0e7d9799eff39da3663, NAME => 'TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='TestLogRolling-testLogRolling', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2 2024-12-05T15:16:42,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741838_1014 (size=64) 2024-12-05T15:16:42,585 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741838_1014 (size=64) 2024-12-05T15:16:42,586 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:42,586 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1681): Closing 7321143ee598c0e7d9799eff39da3663, disabling compactions & flushes 2024-12-05T15:16:42,586 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:42,586 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:42,586 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. after waiting 0 ms 2024-12-05T15:16:42,586 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:42,586 INFO [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:42,586 DEBUG [RegionOpenAndInit-TestLogRolling-testLogRolling-pool-0 {}] regionserver.HRegion(1635): Region close journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:16:42,587 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:16:42,588 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.","families":{"info":[{"qualifier":"regioninfo","vlen":63,"tag":[],"timestamp":"1733411802587"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411802587"}]},"ts":"1733411802587"} 2024-12-05T15:16:42,589 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:16:42,590 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:16:42,590 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411802590"}]},"ts":"1733411802590"} 2024-12-05T15:16:42,592 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRolling, state=ENABLING in hbase:meta 2024-12-05T15:16:42,596 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=7321143ee598c0e7d9799eff39da3663, ASSIGN}] 2024-12-05T15:16:42,597 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=7321143ee598c0e7d9799eff39da3663, ASSIGN 2024-12-05T15:16:42,598 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=7321143ee598c0e7d9799eff39da3663, ASSIGN; state=OFFLINE, location=4ecb0f57b706,43509,1733411800834; forceNewPlan=false, retain=false 2024-12-05T15:16:42,748 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=7321143ee598c0e7d9799eff39da3663, regionState=OPENING, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:42,750 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834}] 2024-12-05T15:16:42,902 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:42,906 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:42,906 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 7321143ee598c0e7d9799eff39da3663, NAME => 'TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:16:42,906 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:42,906 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:16:42,906 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:42,906 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:42,908 INFO [StoreOpener-7321143ee598c0e7d9799eff39da3663-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:42,909 INFO [StoreOpener-7321143ee598c0e7d9799eff39da3663-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 7321143ee598c0e7d9799eff39da3663 columnFamilyName info 2024-12-05T15:16:42,909 DEBUG [StoreOpener-7321143ee598c0e7d9799eff39da3663-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:16:42,910 INFO [StoreOpener-7321143ee598c0e7d9799eff39da3663-1 {}] regionserver.HStore(327): Store=7321143ee598c0e7d9799eff39da3663/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:16:42,910 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:42,911 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:42,912 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:42,914 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:16:42,914 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1102): Opened 7321143ee598c0e7d9799eff39da3663; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=856129, jitterRate=0.08862482011318207}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:16:42,915 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:16:42,916 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663., pid=11, masterSystemTime=1733411802902 2024-12-05T15:16:42,917 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:42,917 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=11}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:42,918 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=7321143ee598c0e7d9799eff39da3663, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:42,921 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2024-12-05T15:16:42,921 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834 in 169 msec 2024-12-05T15:16:42,923 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2024-12-05T15:16:42,923 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=7321143ee598c0e7d9799eff39da3663, ASSIGN in 326 msec 2024-12-05T15:16:42,923 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:16:42,924 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"TestLogRolling-testLogRolling","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411802923"}]},"ts":"1733411802923"} 2024-12-05T15:16:42,925 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=TestLogRolling-testLogRolling, state=ENABLED in hbase:meta 2024-12-05T15:16:42,928 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=TestLogRolling-testLogRolling execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:16:42,930 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=TestLogRolling-testLogRolling in 362 msec 2024-12-05T15:16:43,169 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:43,344 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:44,169 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:44,344 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:44,601 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,601 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,602 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,613 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,613 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,614 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,614 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,614 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,614 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,617 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,617 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,617 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:44,619 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,124 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:16:45,125 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,126 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,126 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,126 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,126 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,127 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,141 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,141 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,141 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,141 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,142 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,142 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,145 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,145 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,145 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,147 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:45,153 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling 2024-12-05T15:16:45,153 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testLogRolling Metrics about Tables on a single HBase RegionServer 2024-12-05T15:16:45,154 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_TestLogRolling-testCompactionRecordDoesntBlockRolling 2024-12-05T15:16:45,170 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:45,345 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:46,170 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:46,346 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:47,171 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:47,346 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:47,483 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2024-12-05T15:16:47,483 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2024-12-05T15:16:47,484 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'TestLogRolling-testLogRolling' 2024-12-05T15:16:48,171 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:48,347 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:48,882 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 817a20dfbf0d1531445f95325bcaf27c, had cached 0 bytes from a total of 23930 2024-12-05T15:16:49,172 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:49,347 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:50,173 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:50,348 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:50,656 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:16:50,657 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,657 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,657 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,657 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,658 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,659 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,673 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,674 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,674 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,674 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,674 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,674 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,677 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,677 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,677 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:50,679 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:16:51,173 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:51,348 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:52,174 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:52,349 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:52,570 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=42659 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2024-12-05T15:16:52,571 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:TestLogRolling-testLogRolling, procId: 9 completed 2024-12-05T15:16:52,573 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 1 regions for table TestLogRolling-testLogRolling 2024-12-05T15:16:52,573 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:16:52,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:16:52,584 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 7321143ee598c0e7d9799eff39da3663 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:16:52,602 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/a0658fd68b9c41368bd3db9cfd284f72 is 1080, key is row0001/info:/1733411812576/Put/seqid=0 2024-12-05T15:16:52,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741839_1015 (size=12509) 2024-12-05T15:16:52,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741839_1015 (size=12509) 2024-12-05T15:16:52,608 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=11 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/a0658fd68b9c41368bd3db9cfd284f72 2024-12-05T15:16:52,608 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-05T15:16:52,608 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] ipc.CallRunner(138): callId: 38 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:56170 deadline: 1733411822608, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834 2024-12-05T15:16:52,614 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/a0658fd68b9c41368bd3db9cfd284f72 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/a0658fd68b9c41368bd3db9cfd284f72 2024-12-05T15:16:52,619 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/a0658fd68b9c41368bd3db9cfd284f72, entries=7, sequenceid=11, filesize=12.2 K 2024-12-05T15:16:52,620 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for 7321143ee598c0e7d9799eff39da3663 in 37ms, sequenceid=11, compaction requested=false 2024-12-05T15:16:52,620 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:16:53,174 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:53,349 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:54,175 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:54,350 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:55,175 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:55,326 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): data stats (chunk size=2097152): current pool size=1, created chunk count=15, reused chunk count=36, reuseRatio=70.59% 2024-12-05T15:16:55,327 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): index stats (chunk size=209715): current pool size=0, created chunk count=0, reused chunk count=0, reuseRatio=0 2024-12-05T15:16:55,350 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:56,176 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:56,351 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:57,176 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:57,352 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:58,177 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:58,352 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:59,177 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:16:59,353 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:00,178 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:00,353 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:01,179 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:01,354 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:01,924 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,925 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,925 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,925 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,925 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,926 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,939 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,940 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,940 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,940 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,940 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,940 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,943 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,943 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,943 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:01,945 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,179 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:02,354 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:02,452 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:17:02,453 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,454 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,454 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,454 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,454 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,455 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,470 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,470 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,470 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,471 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,471 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,471 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,474 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,474 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,474 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,476 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:02,705 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:02,705 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 7321143ee598c0e7d9799eff39da3663 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-05T15:17:02,711 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/9af6983241444f37aa891093f8540681 is 1080, key is row0008/info:/1733411812584/Put/seqid=0 2024-12-05T15:17:02,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741840_1016 (size=29761) 2024-12-05T15:17:02,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741840_1016 (size=29761) 2024-12-05T15:17:02,716 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=37 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/9af6983241444f37aa891093f8540681 2024-12-05T15:17:02,722 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/9af6983241444f37aa891093f8540681 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681 2024-12-05T15:17:02,728 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681, entries=23, sequenceid=37, filesize=29.1 K 2024-12-05T15:17:02,728 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=2.10 KB/2152 for 7321143ee598c0e7d9799eff39da3663 in 23ms, sequenceid=37, compaction requested=false 2024-12-05T15:17:02,729 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:17:02,729 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=41.3 K, sizeToCheck=16.0 K 2024-12-05T15:17:02,729 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:17:02,729 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681 because midkey is the same as first or last row 2024-12-05T15:17:03,180 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:03,355 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:04,180 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:04,356 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:04,714 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:04,714 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 7321143ee598c0e7d9799eff39da3663 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:17:04,718 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/958d011e3057482986633b9fd0aa66ea is 1080, key is row0031/info:/1733411822706/Put/seqid=0 2024-12-05T15:17:04,723 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741841_1017 (size=12509) 2024-12-05T15:17:04,724 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741841_1017 (size=12509) 2024-12-05T15:17:04,724 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=47 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/958d011e3057482986633b9fd0aa66ea 2024-12-05T15:17:04,730 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/958d011e3057482986633b9fd0aa66ea as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/958d011e3057482986633b9fd0aa66ea 2024-12-05T15:17:04,735 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/958d011e3057482986633b9fd0aa66ea, entries=7, sequenceid=47, filesize=12.2 K 2024-12-05T15:17:04,736 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=22.07 KB/22596 for 7321143ee598c0e7d9799eff39da3663 in 22ms, sequenceid=47, compaction requested=true 2024-12-05T15:17:04,736 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:17:04,737 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=53.5 K, sizeToCheck=16.0 K 2024-12-05T15:17:04,737 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:17:04,737 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681 because midkey is the same as first or last row 2024-12-05T15:17:04,737 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:04,737 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 7321143ee598c0e7d9799eff39da3663:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:17:04,737 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:04,737 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:17:04,737 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 7321143ee598c0e7d9799eff39da3663 1/1 column families, dataSize=23.12 KB heapSize=25 KB 2024-12-05T15:17:04,738 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 54779 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:17:04,738 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 7321143ee598c0e7d9799eff39da3663/info is initiating minor compaction (all files) 2024-12-05T15:17:04,738 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 7321143ee598c0e7d9799eff39da3663/info in TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:17:04,738 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/a0658fd68b9c41368bd3db9cfd284f72, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/958d011e3057482986633b9fd0aa66ea] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp, totalSize=53.5 K 2024-12-05T15:17:04,739 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting a0658fd68b9c41368bd3db9cfd284f72, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=11, earliestPutTs=1733411812576 2024-12-05T15:17:04,740 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 9af6983241444f37aa891093f8540681, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=37, earliestPutTs=1733411812584 2024-12-05T15:17:04,740 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 958d011e3057482986633b9fd0aa66ea, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=47, earliestPutTs=1733411822706 2024-12-05T15:17:04,742 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/fd71b8e30e784cfa869b0d88330f6451 is 1080, key is row0038/info:/1733411824714/Put/seqid=0 2024-12-05T15:17:04,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741842_1018 (size=28684) 2024-12-05T15:17:04,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741842_1018 (size=28684) 2024-12-05T15:17:04,754 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=23.12 KB at sequenceid=72 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/fd71b8e30e784cfa869b0d88330f6451 2024-12-05T15:17:04,761 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 7321143ee598c0e7d9799eff39da3663#info#compaction#42 average throughput is 9.49 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:17:04,761 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/fd71b8e30e784cfa869b0d88330f6451 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fd71b8e30e784cfa869b0d88330f6451 2024-12-05T15:17:04,762 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/fe59939287c74c2e9eda6956e6de18de is 1080, key is row0001/info:/1733411812576/Put/seqid=0 2024-12-05T15:17:04,767 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fd71b8e30e784cfa869b0d88330f6451, entries=22, sequenceid=72, filesize=28.0 K 2024-12-05T15:17:04,768 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~23.12 KB/23672, heapSize ~24.98 KB/25584, currentSize=5.25 KB/5380 for 7321143ee598c0e7d9799eff39da3663 in 31ms, sequenceid=72, compaction requested=false 2024-12-05T15:17:04,768 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:17:04,769 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=81.5 K, sizeToCheck=16.0 K 2024-12-05T15:17:04,769 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:17:04,769 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681 because midkey is the same as first or last row 2024-12-05T15:17:04,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741843_1019 (size=44978) 2024-12-05T15:17:04,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741843_1019 (size=44978) 2024-12-05T15:17:04,776 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/fe59939287c74c2e9eda6956e6de18de as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fe59939287c74c2e9eda6956e6de18de 2024-12-05T15:17:04,781 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 7321143ee598c0e7d9799eff39da3663/info of 7321143ee598c0e7d9799eff39da3663 into fe59939287c74c2e9eda6956e6de18de(size=43.9 K), total size for store is 71.9 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:17:04,782 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:17:04,782 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663., storeName=7321143ee598c0e7d9799eff39da3663/info, priority=13, startTime=1733411824737; duration=0sec 2024-12-05T15:17:04,782 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=71.9 K, sizeToCheck=16.0 K 2024-12-05T15:17:04,782 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:17:04,782 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fe59939287c74c2e9eda6956e6de18de because midkey is the same as first or last row 2024-12-05T15:17:04,782 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:04,782 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 7321143ee598c0e7d9799eff39da3663:info 2024-12-05T15:17:05,181 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:05,356 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:06,181 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:06,357 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:06,746 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:06,746 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 7321143ee598c0e7d9799eff39da3663 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:17:06,751 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/1fb1cf9139a74d898160b53b723fe2b1 is 1080, key is row0060/info:/1733411824738/Put/seqid=0 2024-12-05T15:17:06,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741844_1020 (size=12509) 2024-12-05T15:17:06,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741844_1020 (size=12509) 2024-12-05T15:17:06,763 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=83 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/1fb1cf9139a74d898160b53b723fe2b1 2024-12-05T15:17:06,769 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/1fb1cf9139a74d898160b53b723fe2b1 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/1fb1cf9139a74d898160b53b723fe2b1 2024-12-05T15:17:06,771 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-05T15:17:06,771 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] ipc.CallRunner(138): callId: 99 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:56170 deadline: 1733411836771, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:06,775 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/1fb1cf9139a74d898160b53b723fe2b1, entries=7, sequenceid=83, filesize=12.2 K 2024-12-05T15:17:06,776 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for 7321143ee598c0e7d9799eff39da3663 in 29ms, sequenceid=83, compaction requested=true 2024-12-05T15:17:06,776 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:17:06,776 DEBUG [MemStoreFlusher.0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=84.2 K, sizeToCheck=16.0 K 2024-12-05T15:17:06,776 DEBUG [MemStoreFlusher.0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:17:06,776 DEBUG [MemStoreFlusher.0 {}] regionserver.StoreUtils(137): cannot split hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fe59939287c74c2e9eda6956e6de18de because midkey is the same as first or last row 2024-12-05T15:17:06,776 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 7321143ee598c0e7d9799eff39da3663:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:17:06,776 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:06,776 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:17:06,777 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 86171 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:17:06,777 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 7321143ee598c0e7d9799eff39da3663/info is initiating minor compaction (all files) 2024-12-05T15:17:06,777 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 7321143ee598c0e7d9799eff39da3663/info in TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:17:06,777 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fe59939287c74c2e9eda6956e6de18de, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fd71b8e30e784cfa869b0d88330f6451, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/1fb1cf9139a74d898160b53b723fe2b1] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp, totalSize=84.2 K 2024-12-05T15:17:06,777 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting fe59939287c74c2e9eda6956e6de18de, keycount=37, bloomtype=ROW, size=43.9 K, encoding=NONE, compression=NONE, seqNum=47, earliestPutTs=1733411812576 2024-12-05T15:17:06,778 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting fd71b8e30e784cfa869b0d88330f6451, keycount=22, bloomtype=ROW, size=28.0 K, encoding=NONE, compression=NONE, seqNum=72, earliestPutTs=1733411824714 2024-12-05T15:17:06,778 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 1fb1cf9139a74d898160b53b723fe2b1, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=83, earliestPutTs=1733411824738 2024-12-05T15:17:06,789 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 7321143ee598c0e7d9799eff39da3663#info#compaction#44 average throughput is 33.86 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:17:06,790 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/b145471a147d41e587987baecf372419 is 1080, key is row0001/info:/1733411812576/Put/seqid=0 2024-12-05T15:17:06,793 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741845_1021 (size=76455) 2024-12-05T15:17:06,793 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741845_1021 (size=76455) 2024-12-05T15:17:06,801 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/b145471a147d41e587987baecf372419 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419 2024-12-05T15:17:06,807 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 7321143ee598c0e7d9799eff39da3663/info of 7321143ee598c0e7d9799eff39da3663 into b145471a147d41e587987baecf372419(size=74.7 K), total size for store is 74.7 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:17:06,807 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:17:06,807 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663., storeName=7321143ee598c0e7d9799eff39da3663/info, priority=13, startTime=1733411826776; duration=0sec 2024-12-05T15:17:06,807 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.ConstantSizeRegionSplitPolicy(109): Should split because info size=74.7 K, sizeToCheck=16.0 K 2024-12-05T15:17:06,807 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.IncreasingToUpperBoundRegionSplitPolicy(85): regionsWithCommonTable=1 2024-12-05T15:17:06,808 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit(239): Splitting TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663., compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:06,808 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:06,808 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 7321143ee598c0e7d9799eff39da3663:info 2024-12-05T15:17:06,809 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42659 {}] assignment.AssignmentManager(1346): Split request from 4ecb0f57b706,43509,1733411800834, parent={ENCODED => 7321143ee598c0e7d9799eff39da3663, NAME => 'TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.', STARTKEY => '', ENDKEY => ''}, splitKey=row0062 2024-12-05T15:17:06,814 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42659 {}] assignment.SplitTableRegionProcedure(223): Splittable=true state=OPEN, location=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:06,817 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=42659 {}] procedure2.ProcedureExecutor(1098): Stored pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=7321143ee598c0e7d9799eff39da3663, daughterA=ff80d05b17e1b6958441720b0213cd29, daughterB=3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:06,818 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=7321143ee598c0e7d9799eff39da3663, daughterA=ff80d05b17e1b6958441720b0213cd29, daughterB=3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:06,818 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=7321143ee598c0e7d9799eff39da3663, daughterA=ff80d05b17e1b6958441720b0213cd29, daughterB=3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:06,818 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=12, state=RUNNABLE:SPLIT_TABLE_REGION_PREPARE; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=7321143ee598c0e7d9799eff39da3663, daughterA=ff80d05b17e1b6958441720b0213cd29, daughterB=3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:06,824 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=7321143ee598c0e7d9799eff39da3663, UNASSIGN}] 2024-12-05T15:17:06,825 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=7321143ee598c0e7d9799eff39da3663, UNASSIGN 2024-12-05T15:17:06,825 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=7321143ee598c0e7d9799eff39da3663, regionState=CLOSING, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:06,827 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2024-12-05T15:17:06,827 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=14, ppid=13, state=RUNNABLE; CloseRegionProcedure 7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834}] 2024-12-05T15:17:06,982 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:06,983 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] handler.UnassignRegionHandler(124): Close 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:06,983 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2024-12-05T15:17:06,984 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1681): Closing 7321143ee598c0e7d9799eff39da3663, disabling compactions & flushes 2024-12-05T15:17:06,984 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:17:06,984 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:17:06,984 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. after waiting 0 ms 2024-12-05T15:17:06,984 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:17:06,984 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(2837): Flushing 7321143ee598c0e7d9799eff39da3663 1/1 column families, dataSize=23.12 KB heapSize=25 KB 2024-12-05T15:17:06,989 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/45a8e7a068754919a0efbec4b7b18eb0 is 1080, key is row0067/info:/1733411826747/Put/seqid=0 2024-12-05T15:17:06,994 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741846_1022 (size=28684) 2024-12-05T15:17:06,994 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741846_1022 (size=28684) 2024-12-05T15:17:06,994 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=23.12 KB at sequenceid=109 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/45a8e7a068754919a0efbec4b7b18eb0 2024-12-05T15:17:07,000 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/.tmp/info/45a8e7a068754919a0efbec4b7b18eb0 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/45a8e7a068754919a0efbec4b7b18eb0 2024-12-05T15:17:07,005 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/45a8e7a068754919a0efbec4b7b18eb0, entries=22, sequenceid=109, filesize=28.0 K 2024-12-05T15:17:07,006 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(3040): Finished flush of dataSize ~23.12 KB/23672, heapSize ~24.98 KB/25584, currentSize=0 B/0 for 7321143ee598c0e7d9799eff39da3663 in 22ms, sequenceid=109, compaction requested=false 2024-12-05T15:17:07,007 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/a0658fd68b9c41368bd3db9cfd284f72, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fe59939287c74c2e9eda6956e6de18de, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/958d011e3057482986633b9fd0aa66ea, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fd71b8e30e784cfa869b0d88330f6451, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/1fb1cf9139a74d898160b53b723fe2b1] to archive 2024-12-05T15:17:07,008 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-05T15:17:07,010 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/a0658fd68b9c41368bd3db9cfd284f72 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/a0658fd68b9c41368bd3db9cfd284f72 2024-12-05T15:17:07,011 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/9af6983241444f37aa891093f8540681 2024-12-05T15:17:07,012 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fe59939287c74c2e9eda6956e6de18de to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fe59939287c74c2e9eda6956e6de18de 2024-12-05T15:17:07,013 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/958d011e3057482986633b9fd0aa66ea to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/958d011e3057482986633b9fd0aa66ea 2024-12-05T15:17:07,014 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fd71b8e30e784cfa869b0d88330f6451 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/fd71b8e30e784cfa869b0d88330f6451 2024-12-05T15:17:07,015 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/1fb1cf9139a74d898160b53b723fe2b1 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/1fb1cf9139a74d898160b53b723fe2b1 2024-12-05T15:17:07,019 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/recovered.edits/112.seqid, newMaxSeqId=112, maxSeqId=1 2024-12-05T15:17:07,020 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. 2024-12-05T15:17:07,020 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] regionserver.HRegion(1635): Region close journal for 7321143ee598c0e7d9799eff39da3663: 2024-12-05T15:17:07,021 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION, pid=14}] handler.UnassignRegionHandler(170): Closed 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:07,022 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=7321143ee598c0e7d9799eff39da3663, regionState=CLOSED 2024-12-05T15:17:07,025 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=14, resume processing ppid=13 2024-12-05T15:17:07,025 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=14, ppid=13, state=SUCCESS; CloseRegionProcedure 7321143ee598c0e7d9799eff39da3663, server=4ecb0f57b706,43509,1733411800834 in 196 msec 2024-12-05T15:17:07,027 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=13, resume processing ppid=12 2024-12-05T15:17:07,027 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=13, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=7321143ee598c0e7d9799eff39da3663, UNASSIGN in 201 msec 2024-12-05T15:17:07,045 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:17:07,046 INFO [PEWorker-4 {}] assignment.SplitTableRegionProcedure(728): pid=12 splitting 2 storefiles, region=7321143ee598c0e7d9799eff39da3663, threads=2 2024-12-05T15:17:07,047 DEBUG [StoreFileSplitter-pool-0 {}] assignment.SplitTableRegionProcedure(823): pid=12 splitting started for store file: hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/45a8e7a068754919a0efbec4b7b18eb0 for region: 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:07,047 DEBUG [StoreFileSplitter-pool-1 {}] assignment.SplitTableRegionProcedure(823): pid=12 splitting started for store file: hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419 for region: 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:07,055 DEBUG [StoreFileSplitter-pool-0 {}] regionserver.HRegionFileSystem(650): Will create HFileLink file for hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/45a8e7a068754919a0efbec4b7b18eb0, top=true 2024-12-05T15:17:07,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741847_1023 (size=27) 2024-12-05T15:17:07,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741847_1023 (size=27) 2024-12-05T15:17:07,065 INFO [StoreFileSplitter-pool-0 {}] regionserver.HRegionFileSystem(691): Created linkFile:hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0 for child: 3b2c9c51278861a92ceac0eb6ba58da3, parent: 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:07,065 DEBUG [StoreFileSplitter-pool-0 {}] assignment.SplitTableRegionProcedure(834): pid=12 splitting complete for store file: hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/45a8e7a068754919a0efbec4b7b18eb0 for region: 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:07,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741848_1024 (size=27) 2024-12-05T15:17:07,075 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741848_1024 (size=27) 2024-12-05T15:17:07,075 DEBUG [StoreFileSplitter-pool-1 {}] assignment.SplitTableRegionProcedure(834): pid=12 splitting complete for store file: hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419 for region: 7321143ee598c0e7d9799eff39da3663 2024-12-05T15:17:07,075 DEBUG [PEWorker-4 {}] assignment.SplitTableRegionProcedure(802): pid=12 split storefiles for region 7321143ee598c0e7d9799eff39da3663 Daughter A: [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663] storefiles, Daughter B: [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663] storefiles. 2024-12-05T15:17:07,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741849_1025 (size=71) 2024-12-05T15:17:07,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741849_1025 (size=71) 2024-12-05T15:17:07,085 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:17:07,095 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741850_1026 (size=71) 2024-12-05T15:17:07,095 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741850_1026 (size=71) 2024-12-05T15:17:07,097 DEBUG [PEWorker-4 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:17:07,105 DEBUG [PEWorker-4 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/recovered.edits/112.seqid, newMaxSeqId=112, maxSeqId=-1 2024-12-05T15:17:07,106 DEBUG [PEWorker-4 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/recovered.edits/112.seqid, newMaxSeqId=112, maxSeqId=-1 2024-12-05T15:17:07,108 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663.","families":{"info":[{"qualifier":"regioninfo","vlen":63,"tag":[],"timestamp":"1733411827108"},{"qualifier":"splitA","vlen":70,"tag":[],"timestamp":"1733411827108"},{"qualifier":"splitB","vlen":70,"tag":[],"timestamp":"1733411827108"}]},"ts":"1733411827108"} 2024-12-05T15:17:07,108 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29.","families":{"info":[{"qualifier":"regioninfo","vlen":70,"tag":[],"timestamp":"1733411827108"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411827108"},{"qualifier":"seqnumDuringOpen","vlen":8,"tag":[],"timestamp":"1733411827108"}]},"ts":"1733411827108"} 2024-12-05T15:17:07,109 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":3,"row":"TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.","families":{"info":[{"qualifier":"regioninfo","vlen":70,"tag":[],"timestamp":"1733411827108"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411827108"},{"qualifier":"seqnumDuringOpen","vlen":8,"tag":[],"timestamp":"1733411827108"}]},"ts":"1733411827108"} 2024-12-05T15:17:07,138 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=43509 {}] regionserver.HRegion(8581): Flush requested on 1588230740 2024-12-05T15:17:07,138 DEBUG [MemStoreFlusher.0 {}] regionserver.FlushAllLargeStoresPolicy(69): Since none of the CFs were above the size, flushing all. 2024-12-05T15:17:07,139 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=4.75 KB heapSize=8.29 KB 2024-12-05T15:17:07,143 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=15, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=ff80d05b17e1b6958441720b0213cd29, ASSIGN}, {pid=16, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=3b2c9c51278861a92ceac0eb6ba58da3, ASSIGN}] 2024-12-05T15:17:07,144 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=16, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=3b2c9c51278861a92ceac0eb6ba58da3, ASSIGN 2024-12-05T15:17:07,144 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=15, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=ff80d05b17e1b6958441720b0213cd29, ASSIGN 2024-12-05T15:17:07,145 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=16, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=3b2c9c51278861a92ceac0eb6ba58da3, ASSIGN; state=SPLITTING_NEW, location=4ecb0f57b706,43509,1733411800834; forceNewPlan=false, retain=false 2024-12-05T15:17:07,145 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=15, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=ff80d05b17e1b6958441720b0213cd29, ASSIGN; state=SPLITTING_NEW, location=4ecb0f57b706,43509,1733411800834; forceNewPlan=false, retain=false 2024-12-05T15:17:07,155 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/info/b8216a2fac3e4e07a2735fce2fb1c40d is 193, key is TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3./info:regioninfo/1733411827108/Put/seqid=0 2024-12-05T15:17:07,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741851_1027 (size=9423) 2024-12-05T15:17:07,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741851_1027 (size=9423) 2024-12-05T15:17:07,160 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=4.54 KB at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/info/b8216a2fac3e4e07a2735fce2fb1c40d 2024-12-05T15:17:07,182 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:07,187 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/table/349294ba7f0a4a02821f44834e2e4bf7 is 65, key is TestLogRolling-testLogRolling/table:state/1733411802923/Put/seqid=0 2024-12-05T15:17:07,192 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741852_1028 (size=5412) 2024-12-05T15:17:07,192 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741852_1028 (size=5412) 2024-12-05T15:17:07,193 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=216 B at sequenceid=17 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/table/349294ba7f0a4a02821f44834e2e4bf7 2024-12-05T15:17:07,198 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/info/b8216a2fac3e4e07a2735fce2fb1c40d as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/info/b8216a2fac3e4e07a2735fce2fb1c40d 2024-12-05T15:17:07,204 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/info/b8216a2fac3e4e07a2735fce2fb1c40d, entries=29, sequenceid=17, filesize=9.2 K 2024-12-05T15:17:07,205 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/table/349294ba7f0a4a02821f44834e2e4bf7 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/table/349294ba7f0a4a02821f44834e2e4bf7 2024-12-05T15:17:07,210 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/table/349294ba7f0a4a02821f44834e2e4bf7, entries=4, sequenceid=17, filesize=5.3 K 2024-12-05T15:17:07,211 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~4.75 KB/4869, heapSize ~8.01 KB/8200, currentSize=0 B/0 for 1588230740 in 73ms, sequenceid=17, compaction requested=false 2024-12-05T15:17:07,211 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 1588230740: 2024-12-05T15:17:07,295 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=15 updating hbase:meta row=ff80d05b17e1b6958441720b0213cd29, regionState=OPENING, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:07,295 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=16 updating hbase:meta row=3b2c9c51278861a92ceac0eb6ba58da3, regionState=OPENING, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:07,297 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=17, ppid=15, state=RUNNABLE; OpenRegionProcedure ff80d05b17e1b6958441720b0213cd29, server=4ecb0f57b706,43509,1733411800834}] 2024-12-05T15:17:07,298 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=18, ppid=16, state=RUNNABLE; OpenRegionProcedure 3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834}] 2024-12-05T15:17:07,357 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:07,450 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:07,453 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:17:07,453 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(7285): Opening region: {ENCODED => ff80d05b17e1b6958441720b0213cd29, NAME => 'TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29.', STARTKEY => '', ENDKEY => 'row0062'} 2024-12-05T15:17:07,454 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:17:07,454 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:17:07,454 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(7327): checking encryption for ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:17:07,454 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(7330): checking classloading for ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:17:07,455 INFO [StoreOpener-ff80d05b17e1b6958441720b0213cd29-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:17:07,456 INFO [StoreOpener-ff80d05b17e1b6958441720b0213cd29-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region ff80d05b17e1b6958441720b0213cd29 columnFamilyName info 2024-12-05T15:17:07,456 DEBUG [StoreOpener-ff80d05b17e1b6958441720b0213cd29-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:17:07,468 DEBUG [StoreOpener-ff80d05b17e1b6958441720b0213cd29-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663->hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419-bottom 2024-12-05T15:17:07,468 INFO [StoreOpener-ff80d05b17e1b6958441720b0213cd29-1 {}] regionserver.HStore(327): Store=ff80d05b17e1b6958441720b0213cd29/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:17:07,469 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:17:07,470 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:17:07,472 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(1085): writing seq id for ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:17:07,473 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(1102): Opened ff80d05b17e1b6958441720b0213cd29; next sequenceid=113; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=868111, jitterRate=0.10386060178279877}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:17:07,473 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegion(1001): Region open journal for ff80d05b17e1b6958441720b0213cd29: 2024-12-05T15:17:07,474 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29., pid=17, masterSystemTime=1733411827450 2024-12-05T15:17:07,474 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.CompactSplit(403): Add compact mark for store ff80d05b17e1b6958441720b0213cd29:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:17:07,474 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: Opening Region; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:07,475 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 1 store files, 0 compacting, 1 eligible, 16 blocking 2024-12-05T15:17:07,475 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1526): Keeping/Overriding Compaction request priority to -2147482648 for CF info since it belongs to recently split daughter region TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:17:07,475 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): ff80d05b17e1b6958441720b0213cd29/info is initiating minor compaction (all files) 2024-12-05T15:17:07,475 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of ff80d05b17e1b6958441720b0213cd29/info in TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:17:07,475 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663->hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419-bottom] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/.tmp, totalSize=74.7 K 2024-12-05T15:17:07,476 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663, keycount=33, bloomtype=ROW, size=74.7 K, encoding=NONE, compression=NONE, seqNum=83, earliestPutTs=1733411812576 2024-12-05T15:17:07,476 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:17:07,476 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=17}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:17:07,477 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] handler.AssignRegionHandler(135): Open TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:17:07,477 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7285): Opening region: {ENCODED => 3b2c9c51278861a92ceac0eb6ba58da3, NAME => 'TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.', STARTKEY => 'row0062', ENDKEY => ''} 2024-12-05T15:17:07,477 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=15 updating hbase:meta row=ff80d05b17e1b6958441720b0213cd29, regionState=OPEN, openSeqNum=113, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:07,477 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table TestLogRolling-testLogRolling 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:07,477 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(894): Instantiated TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:17:07,477 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7327): checking encryption for 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:07,477 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(7330): checking classloading for 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:07,478 INFO [StoreOpener-3b2c9c51278861a92ceac0eb6ba58da3-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:07,479 INFO [StoreOpener-3b2c9c51278861a92ceac0eb6ba58da3-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 3b2c9c51278861a92ceac0eb6ba58da3 columnFamilyName info 2024-12-05T15:17:07,479 DEBUG [StoreOpener-3b2c9c51278861a92ceac0eb6ba58da3-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:17:07,480 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=17, resume processing ppid=15 2024-12-05T15:17:07,481 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=17, ppid=15, state=SUCCESS; OpenRegionProcedure ff80d05b17e1b6958441720b0213cd29, server=4ecb0f57b706,43509,1733411800834 in 181 msec 2024-12-05T15:17:07,482 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=15, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=ff80d05b17e1b6958441720b0213cd29, ASSIGN in 338 msec 2024-12-05T15:17:07,487 DEBUG [StoreOpener-3b2c9c51278861a92ceac0eb6ba58da3-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0 2024-12-05T15:17:07,491 DEBUG [StoreOpener-3b2c9c51278861a92ceac0eb6ba58da3-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663->hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419-top 2024-12-05T15:17:07,491 INFO [StoreOpener-3b2c9c51278861a92ceac0eb6ba58da3-1 {}] regionserver.HStore(327): Store=3b2c9c51278861a92ceac0eb6ba58da3/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:17:07,492 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:07,493 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:07,495 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1085): writing seq id for 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:07,496 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1102): Opened 3b2c9c51278861a92ceac0eb6ba58da3; next sequenceid=113; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=712445, jitterRate=-0.09408026933670044}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:17:07,496 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegion(1001): Region open journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:07,497 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegionServer(2601): Post open deploy tasks for TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., pid=18, masterSystemTime=1733411827450 2024-12-05T15:17:07,497 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.CompactSplit(403): Add compact mark for store 3b2c9c51278861a92ceac0eb6ba58da3:info, priority=-2147483648, current under compaction store size is 2 2024-12-05T15:17:07,497 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: Opening Region; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:07,497 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): ff80d05b17e1b6958441720b0213cd29#info#compaction#48 average throughput is 20.87 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:17:07,497 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 2 store files, 0 compacting, 2 eligible, 16 blocking 2024-12-05T15:17:07,498 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/.tmp/info/c182b911c97d4d8ea15da73039f826e4 is 1080, key is row0001/info:/1733411812576/Put/seqid=0 2024-12-05T15:17:07,498 INFO [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.HStore(1526): Keeping/Overriding Compaction request priority to -2147482648 for CF info since it belongs to recently split daughter region TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:17:07,498 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.HStore(1540): 3b2c9c51278861a92ceac0eb6ba58da3/info is initiating minor compaction (all files) 2024-12-05T15:17:07,498 INFO [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 3b2c9c51278861a92ceac0eb6ba58da3/info in TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:17:07,498 INFO [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663->hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419-top, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp, totalSize=102.7 K 2024-12-05T15:17:07,499 DEBUG [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] regionserver.HRegionServer(2628): Finished post open deploy task for TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:17:07,499 INFO [RS_OPEN_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_REGION, pid=18}] handler.AssignRegionHandler(164): Opened TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:17:07,499 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] compactions.Compactor(224): Compacting b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663, keycount=33, bloomtype=ROW, size=74.7 K, encoding=NONE, compression=NONE, seqNum=84, earliestPutTs=1733411812576 2024-12-05T15:17:07,499 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] compactions.Compactor(224): Compacting TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0, keycount=22, bloomtype=ROW, size=28.0 K, encoding=NONE, compression=NONE, seqNum=109, earliestPutTs=1733411826747 2024-12-05T15:17:07,499 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=16 updating hbase:meta row=3b2c9c51278861a92ceac0eb6ba58da3, regionState=OPEN, openSeqNum=113, regionLocation=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:07,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741853_1029 (size=70862) 2024-12-05T15:17:07,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741853_1029 (size=70862) 2024-12-05T15:17:07,506 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=18, resume processing ppid=16 2024-12-05T15:17:07,507 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=18, ppid=16, state=SUCCESS; OpenRegionProcedure 3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 in 206 msec 2024-12-05T15:17:07,511 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=16, resume processing ppid=12 2024-12-05T15:17:07,511 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=16, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=TestLogRolling-testLogRolling, region=3b2c9c51278861a92ceac0eb6ba58da3, ASSIGN in 364 msec 2024-12-05T15:17:07,512 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/.tmp/info/c182b911c97d4d8ea15da73039f826e4 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/info/c182b911c97d4d8ea15da73039f826e4 2024-12-05T15:17:07,512 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=12, state=SUCCESS; SplitTableRegionProcedure table=TestLogRolling-testLogRolling, parent=7321143ee598c0e7d9799eff39da3663, daughterA=ff80d05b17e1b6958441720b0213cd29, daughterB=3b2c9c51278861a92ceac0eb6ba58da3 in 697 msec 2024-12-05T15:17:07,518 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 1 (all) file(s) in ff80d05b17e1b6958441720b0213cd29/info of ff80d05b17e1b6958441720b0213cd29 into c182b911c97d4d8ea15da73039f826e4(size=69.2 K), total size for store is 69.2 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:17:07,518 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for ff80d05b17e1b6958441720b0213cd29: 2024-12-05T15:17:07,518 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29., storeName=ff80d05b17e1b6958441720b0213cd29/info, priority=15, startTime=1733411827474; duration=0sec 2024-12-05T15:17:07,518 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:07,518 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: ff80d05b17e1b6958441720b0213cd29:info 2024-12-05T15:17:07,529 INFO [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] throttle.PressureAwareThroughputController(145): 3b2c9c51278861a92ceac0eb6ba58da3#info#compaction#49 average throughput is 13.85 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:17:07,530 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/44bce3b7c8d3499e8962867a05bf579e is 1080, key is row0062/info:/1733411824740/Put/seqid=0 2024-12-05T15:17:07,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741854_1030 (size=34267) 2024-12-05T15:17:07,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741854_1030 (size=34267) 2024-12-05T15:17:07,540 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/44bce3b7c8d3499e8962867a05bf579e as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/44bce3b7c8d3499e8962867a05bf579e 2024-12-05T15:17:07,545 INFO [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 2 (all) file(s) in 3b2c9c51278861a92ceac0eb6ba58da3/info of 3b2c9c51278861a92ceac0eb6ba58da3 into 44bce3b7c8d3499e8962867a05bf579e(size=33.5 K), total size for store is 33.5 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:17:07,545 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:07,545 INFO [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., storeName=3b2c9c51278861a92ceac0eb6ba58da3/info, priority=14, startTime=1733411827497; duration=0sec 2024-12-05T15:17:07,546 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:07,546 DEBUG [RS:0;4ecb0f57b706:43509-longCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 3b2c9c51278861a92ceac0eb6ba58da3:info 2024-12-05T15:17:08,182 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:08,358 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:09,183 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:09,358 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:10,183 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:10,359 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:10,776 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:17:11,184 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:11,359 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:12,184 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:12,360 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:12,522 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-datanode.properties,hadoop-metrics2.properties 2024-12-05T15:17:12,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,523 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,525 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,539 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,540 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,540 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,540 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,540 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,540 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,543 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,543 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,543 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:12,545 WARN [HBase-Metrics2-1 {}] impl.FsDatasetImpl(779): Exception thrown while metric collection. Exception : Cannot invoke "java.util.Map.values()" because "this.executors" is null 2024-12-05T15:17:13,185 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:13,360 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:13,467 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:17:13,469 INFO [RS-EventLoopGroup-12-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:49512, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.5 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:17:14,185 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:14,361 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:15,186 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:15,361 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:16,186 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:16,362 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:16,810 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] ipc.CallRunner(138): callId: 101 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:56170 deadline: 1733411846810, exception=org.apache.hadoop.hbase.NotServingRegionException: TestLogRolling-testLogRolling,,1733411802566.7321143ee598c0e7d9799eff39da3663. is not online on 4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:17,187 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:17,362 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:18,187 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:18,363 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:19,188 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:19,363 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:20,189 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:20,364 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:21,189 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:21,365 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:22,190 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:22,365 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:22,475 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2024-12-05T15:17:22,475 INFO [master/4ecb0f57b706:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2024-12-05T15:17:23,190 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:23,366 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:24,191 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:24,366 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:25,191 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:25,367 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:26,192 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:26,367 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:26,984 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 1588230740, had cached 0 bytes from a total of 14835 2024-12-05T15:17:27,192 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:27,368 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:28,193 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:28,368 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:29,193 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:29,369 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:30,194 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:30,369 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:31,194 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:31,370 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:32,195 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:32,370 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:33,195 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:33,196 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=3 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 after 196117ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor237.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:17:33,371 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:33,883 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 817a20dfbf0d1531445f95325bcaf27c, had cached 0 bytes from a total of 23930 2024-12-05T15:17:34,196 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:34,371 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:35,197 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:35,372 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:36,197 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:36,372 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:36,918 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:36,918 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:17:36,923 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/229055645d304b24865458b85f2d6c17 is 1080, key is row0089/info:/1733411856911/Put/seqid=0 2024-12-05T15:17:36,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741855_1031 (size=12509) 2024-12-05T15:17:36,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741855_1031 (size=12509) 2024-12-05T15:17:36,928 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=123 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/229055645d304b24865458b85f2d6c17 2024-12-05T15:17:36,933 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/229055645d304b24865458b85f2d6c17 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/229055645d304b24865458b85f2d6c17 2024-12-05T15:17:36,938 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/229055645d304b24865458b85f2d6c17, entries=7, sequenceid=123, filesize=12.2 K 2024-12-05T15:17:36,939 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=1.05 KB/1076 for 3b2c9c51278861a92ceac0eb6ba58da3 in 21ms, sequenceid=123, compaction requested=false 2024-12-05T15:17:36,939 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:37,198 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:37,373 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:38,198 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:38,373 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:38,927 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:38,927 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:17:38,931 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/432c031e67924a878efbec9c8e093035 is 1080, key is row0096/info:/1733411856919/Put/seqid=0 2024-12-05T15:17:38,936 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741856_1032 (size=12514) 2024-12-05T15:17:38,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741856_1032 (size=12514) 2024-12-05T15:17:38,937 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=133 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/432c031e67924a878efbec9c8e093035 2024-12-05T15:17:38,944 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/432c031e67924a878efbec9c8e093035 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/432c031e67924a878efbec9c8e093035 2024-12-05T15:17:38,949 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/432c031e67924a878efbec9c8e093035, entries=7, sequenceid=133, filesize=12.2 K 2024-12-05T15:17:38,950 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=21.02 KB/21520 for 3b2c9c51278861a92ceac0eb6ba58da3 in 23ms, sequenceid=133, compaction requested=true 2024-12-05T15:17:38,950 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:38,950 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 3b2c9c51278861a92ceac0eb6ba58da3:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:17:38,950 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:38,950 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:17:38,951 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:38,951 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=22.07 KB heapSize=23.88 KB 2024-12-05T15:17:38,951 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 59290 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:17:38,952 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 3b2c9c51278861a92ceac0eb6ba58da3/info is initiating minor compaction (all files) 2024-12-05T15:17:38,952 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 3b2c9c51278861a92ceac0eb6ba58da3/info in TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:17:38,952 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/44bce3b7c8d3499e8962867a05bf579e, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/229055645d304b24865458b85f2d6c17, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/432c031e67924a878efbec9c8e093035] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp, totalSize=57.9 K 2024-12-05T15:17:38,952 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 44bce3b7c8d3499e8962867a05bf579e, keycount=27, bloomtype=ROW, size=33.5 K, encoding=NONE, compression=NONE, seqNum=109, earliestPutTs=1733411824740 2024-12-05T15:17:38,953 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 229055645d304b24865458b85f2d6c17, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=123, earliestPutTs=1733411856911 2024-12-05T15:17:38,953 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 432c031e67924a878efbec9c8e093035, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=133, earliestPutTs=1733411856919 2024-12-05T15:17:38,978 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/741e3d924335466f9bdbf87436ef2b1c is 1080, key is row0103/info:/1733411858927/Put/seqid=0 2024-12-05T15:17:38,985 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741857_1033 (size=27628) 2024-12-05T15:17:38,985 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741857_1033 (size=27628) 2024-12-05T15:17:38,986 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=22.07 KB at sequenceid=157 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/741e3d924335466f9bdbf87436ef2b1c 2024-12-05T15:17:38,988 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 3b2c9c51278861a92ceac0eb6ba58da3#info#compaction#53 average throughput is 42.07 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:17:38,989 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/c97b24774b9d442ab6869227aad0989a is 1080, key is row0062/info:/1733411824740/Put/seqid=0 2024-12-05T15:17:38,991 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/741e3d924335466f9bdbf87436ef2b1c as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/741e3d924335466f9bdbf87436ef2b1c 2024-12-05T15:17:38,993 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741858_1034 (size=49456) 2024-12-05T15:17:38,993 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741858_1034 (size=49456) 2024-12-05T15:17:38,996 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/741e3d924335466f9bdbf87436ef2b1c, entries=21, sequenceid=157, filesize=27.0 K 2024-12-05T15:17:38,997 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~22.07 KB/22596, heapSize ~23.86 KB/24432, currentSize=5.25 KB/5380 for 3b2c9c51278861a92ceac0eb6ba58da3 in 46ms, sequenceid=157, compaction requested=false 2024-12-05T15:17:38,998 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:38,999 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/c97b24774b9d442ab6869227aad0989a as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/c97b24774b9d442ab6869227aad0989a 2024-12-05T15:17:39,005 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 3b2c9c51278861a92ceac0eb6ba58da3/info of 3b2c9c51278861a92ceac0eb6ba58da3 into c97b24774b9d442ab6869227aad0989a(size=48.3 K), total size for store is 75.3 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:17:39,005 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:39,005 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., storeName=3b2c9c51278861a92ceac0eb6ba58da3/info, priority=13, startTime=1733411858950; duration=0sec 2024-12-05T15:17:39,005 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:39,005 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 3b2c9c51278861a92ceac0eb6ba58da3:info 2024-12-05T15:17:39,199 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:39,374 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:40,200 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:40,375 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:40,776 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:17:40,984 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:40,984 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:17:40,988 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/9490ecd40b03448a86d7d0a1fc54ce81 is 1080, key is row0124/info:/1733411858951/Put/seqid=0 2024-12-05T15:17:40,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741859_1035 (size=12516) 2024-12-05T15:17:40,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741859_1035 (size=12516) 2024-12-05T15:17:40,998 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=168 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/9490ecd40b03448a86d7d0a1fc54ce81 2024-12-05T15:17:41,005 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/9490ecd40b03448a86d7d0a1fc54ce81 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/9490ecd40b03448a86d7d0a1fc54ce81 2024-12-05T15:17:41,009 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-05T15:17:41,009 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] ipc.CallRunner(138): callId: 169 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:56170 deadline: 1733411871009, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:41,011 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/9490ecd40b03448a86d7d0a1fc54ce81, entries=7, sequenceid=168, filesize=12.2 K 2024-12-05T15:17:41,012 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for 3b2c9c51278861a92ceac0eb6ba58da3 in 28ms, sequenceid=168, compaction requested=true 2024-12-05T15:17:41,012 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:41,012 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 3b2c9c51278861a92ceac0eb6ba58da3:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:17:41,012 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:41,012 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:17:41,013 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 89600 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:17:41,013 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 3b2c9c51278861a92ceac0eb6ba58da3/info is initiating minor compaction (all files) 2024-12-05T15:17:41,013 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 3b2c9c51278861a92ceac0eb6ba58da3/info in TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:17:41,014 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/c97b24774b9d442ab6869227aad0989a, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/741e3d924335466f9bdbf87436ef2b1c, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/9490ecd40b03448a86d7d0a1fc54ce81] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp, totalSize=87.5 K 2024-12-05T15:17:41,014 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting c97b24774b9d442ab6869227aad0989a, keycount=41, bloomtype=ROW, size=48.3 K, encoding=NONE, compression=NONE, seqNum=133, earliestPutTs=1733411824740 2024-12-05T15:17:41,014 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 741e3d924335466f9bdbf87436ef2b1c, keycount=21, bloomtype=ROW, size=27.0 K, encoding=NONE, compression=NONE, seqNum=157, earliestPutTs=1733411858927 2024-12-05T15:17:41,015 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 9490ecd40b03448a86d7d0a1fc54ce81, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=168, earliestPutTs=1733411858951 2024-12-05T15:17:41,025 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 3b2c9c51278861a92ceac0eb6ba58da3#info#compaction#55 average throughput is 35.40 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:17:41,025 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/8af7a398f44941eb8c188f298510acbd is 1080, key is row0062/info:/1733411824740/Put/seqid=0 2024-12-05T15:17:41,029 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741860_1036 (size=79883) 2024-12-05T15:17:41,029 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741860_1036 (size=79883) 2024-12-05T15:17:41,035 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/8af7a398f44941eb8c188f298510acbd as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/8af7a398f44941eb8c188f298510acbd 2024-12-05T15:17:41,041 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 3b2c9c51278861a92ceac0eb6ba58da3/info of 3b2c9c51278861a92ceac0eb6ba58da3 into 8af7a398f44941eb8c188f298510acbd(size=78.0 K), total size for store is 78.0 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:17:41,041 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:41,041 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., storeName=3b2c9c51278861a92ceac0eb6ba58da3/info, priority=13, startTime=1733411861012; duration=0sec 2024-12-05T15:17:41,042 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:17:41,042 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 3b2c9c51278861a92ceac0eb6ba58da3:info 2024-12-05T15:17:41,200 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:41,375 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:42,166 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2024-12-05T15:17:42,167 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 14ad900f35513e0519981bcd843ee526 changed from -1.0 to 0.0, refreshing cache 2024-12-05T15:17:42,201 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:42,376 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:43,201 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:43,376 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:44,202 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:44,377 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:45,202 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:45,377 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:46,203 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:46,378 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:47,203 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:47,378 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:48,204 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:48,379 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:49,204 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:49,379 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:50,205 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:50,380 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:51,013 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:17:51,013 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-05T15:17:51,018 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/31f7c14f8323441b9691b989810a14bd is 1080, key is row0131/info:/1733411860985/Put/seqid=0 2024-12-05T15:17:51,020 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-05T15:17:51,021 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] ipc.CallRunner(138): callId: 178 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:56170 deadline: 1733411881020, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 2024-12-05T15:17:51,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741861_1037 (size=29784) 2024-12-05T15:17:51,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741861_1037 (size=29784) 2024-12-05T15:17:51,029 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=195 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/31f7c14f8323441b9691b989810a14bd 2024-12-05T15:17:51,035 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/31f7c14f8323441b9691b989810a14bd as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/31f7c14f8323441b9691b989810a14bd 2024-12-05T15:17:51,040 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/31f7c14f8323441b9691b989810a14bd, entries=23, sequenceid=195, filesize=29.1 K 2024-12-05T15:17:51,041 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=6.30 KB/6456 for 3b2c9c51278861a92ceac0eb6ba58da3 in 28ms, sequenceid=195, compaction requested=false 2024-12-05T15:17:51,041 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:17:51,205 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:51,380 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:51,381 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=3 on file=hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta after 196115ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.GeneratedMethodAccessor237.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2024-12-05T15:17:52,206 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:52,381 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:52,454 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region ff80d05b17e1b6958441720b0213cd29, had cached 0 bytes from a total of 70862 2024-12-05T15:17:52,477 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 3b2c9c51278861a92ceac0eb6ba58da3, had cached 0 bytes from a total of 109667 2024-12-05T15:17:53,206 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:53,382 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:54,207 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:54,382 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:55,207 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:55,383 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:56,208 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:56,383 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:57,208 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:57,384 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:58,209 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:58,384 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:59,209 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:17:59,385 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:00,210 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:00,385 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:01,119 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:01,120 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:18:01,125 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/eae7dc227fec41dfb2624feb2d94190c is 1080, key is row0154/info:/1733411871014/Put/seqid=0 2024-12-05T15:18:01,130 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741862_1038 (size=12516) 2024-12-05T15:18:01,130 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741862_1038 (size=12516) 2024-12-05T15:18:01,130 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=205 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/eae7dc227fec41dfb2624feb2d94190c 2024-12-05T15:18:01,137 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/eae7dc227fec41dfb2624feb2d94190c as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/eae7dc227fec41dfb2624feb2d94190c 2024-12-05T15:18:01,142 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/eae7dc227fec41dfb2624feb2d94190c, entries=7, sequenceid=205, filesize=12.2 K 2024-12-05T15:18:01,143 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=0 B/0 for 3b2c9c51278861a92ceac0eb6ba58da3 in 23ms, sequenceid=205, compaction requested=true 2024-12-05T15:18:01,143 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:01,143 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 3b2c9c51278861a92ceac0eb6ba58da3:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:18:01,143 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:01,143 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:18:01,144 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 122183 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:18:01,144 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 3b2c9c51278861a92ceac0eb6ba58da3/info is initiating minor compaction (all files) 2024-12-05T15:18:01,144 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 3b2c9c51278861a92ceac0eb6ba58da3/info in TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:01,144 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/8af7a398f44941eb8c188f298510acbd, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/31f7c14f8323441b9691b989810a14bd, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/eae7dc227fec41dfb2624feb2d94190c] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp, totalSize=119.3 K 2024-12-05T15:18:01,145 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 8af7a398f44941eb8c188f298510acbd, keycount=69, bloomtype=ROW, size=78.0 K, encoding=NONE, compression=NONE, seqNum=168, earliestPutTs=1733411824740 2024-12-05T15:18:01,145 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 31f7c14f8323441b9691b989810a14bd, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=195, earliestPutTs=1733411860985 2024-12-05T15:18:01,145 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting eae7dc227fec41dfb2624feb2d94190c, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=205, earliestPutTs=1733411871014 2024-12-05T15:18:01,160 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 3b2c9c51278861a92ceac0eb6ba58da3#info#compaction#58 average throughput is 33.86 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:18:01,161 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/e39056b86c174aa7b20f150676701d79 is 1080, key is row0062/info:/1733411824740/Put/seqid=0 2024-12-05T15:18:01,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741863_1039 (size=112333) 2024-12-05T15:18:01,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741863_1039 (size=112333) 2024-12-05T15:18:01,172 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/e39056b86c174aa7b20f150676701d79 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e39056b86c174aa7b20f150676701d79 2024-12-05T15:18:01,178 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 3b2c9c51278861a92ceac0eb6ba58da3/info of 3b2c9c51278861a92ceac0eb6ba58da3 into e39056b86c174aa7b20f150676701d79(size=109.7 K), total size for store is 109.7 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:18:01,178 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:01,178 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., storeName=3b2c9c51278861a92ceac0eb6ba58da3/info, priority=13, startTime=1733411881143; duration=0sec 2024-12-05T15:18:01,178 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:01,178 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 3b2c9c51278861a92ceac0eb6ba58da3:info 2024-12-05T15:18:01,210 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:01,386 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:02,211 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:02,386 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:03,127 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:03,127 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:18:03,132 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/e02585c3910048d084020c402d4deb86 is 1080, key is row0161/info:/1733411883121/Put/seqid=0 2024-12-05T15:18:03,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741864_1040 (size=12516) 2024-12-05T15:18:03,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741864_1040 (size=12516) 2024-12-05T15:18:03,138 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=216 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/e02585c3910048d084020c402d4deb86 2024-12-05T15:18:03,144 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/e02585c3910048d084020c402d4deb86 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e02585c3910048d084020c402d4deb86 2024-12-05T15:18:03,150 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e02585c3910048d084020c402d4deb86, entries=7, sequenceid=216, filesize=12.2 K 2024-12-05T15:18:03,151 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for 3b2c9c51278861a92ceac0eb6ba58da3 in 23ms, sequenceid=216, compaction requested=false 2024-12-05T15:18:03,151 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:03,151 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:03,151 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-05T15:18:03,155 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/aae3b33e969b4f9bbea8bbbf689f5413 is 1080, key is row0168/info:/1733411883128/Put/seqid=0 2024-12-05T15:18:03,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741865_1041 (size=29784) 2024-12-05T15:18:03,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741865_1041 (size=29784) 2024-12-05T15:18:03,161 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=242 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/aae3b33e969b4f9bbea8bbbf689f5413 2024-12-05T15:18:03,166 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/aae3b33e969b4f9bbea8bbbf689f5413 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/aae3b33e969b4f9bbea8bbbf689f5413 2024-12-05T15:18:03,171 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/aae3b33e969b4f9bbea8bbbf689f5413, entries=23, sequenceid=242, filesize=29.1 K 2024-12-05T15:18:03,172 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=2.10 KB/2152 for 3b2c9c51278861a92ceac0eb6ba58da3 in 21ms, sequenceid=242, compaction requested=true 2024-12-05T15:18:03,172 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:03,172 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 3b2c9c51278861a92ceac0eb6ba58da3:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:18:03,172 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:03,172 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:18:03,173 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 154633 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:18:03,173 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 3b2c9c51278861a92ceac0eb6ba58da3/info is initiating minor compaction (all files) 2024-12-05T15:18:03,173 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 3b2c9c51278861a92ceac0eb6ba58da3/info in TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:03,173 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e39056b86c174aa7b20f150676701d79, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e02585c3910048d084020c402d4deb86, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/aae3b33e969b4f9bbea8bbbf689f5413] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp, totalSize=151.0 K 2024-12-05T15:18:03,173 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting e39056b86c174aa7b20f150676701d79, keycount=99, bloomtype=ROW, size=109.7 K, encoding=NONE, compression=NONE, seqNum=205, earliestPutTs=1733411824740 2024-12-05T15:18:03,174 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting e02585c3910048d084020c402d4deb86, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=216, earliestPutTs=1733411883121 2024-12-05T15:18:03,174 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting aae3b33e969b4f9bbea8bbbf689f5413, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=242, earliestPutTs=1733411883128 2024-12-05T15:18:03,185 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 3b2c9c51278861a92ceac0eb6ba58da3#info#compaction#61 average throughput is 44.12 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:18:03,185 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/6863d90e88074e528626a6a5e511f7dd is 1080, key is row0062/info:/1733411824740/Put/seqid=0 2024-12-05T15:18:03,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741866_1042 (size=144980) 2024-12-05T15:18:03,189 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741866_1042 (size=144980) 2024-12-05T15:18:03,195 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/6863d90e88074e528626a6a5e511f7dd as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/6863d90e88074e528626a6a5e511f7dd 2024-12-05T15:18:03,201 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 3b2c9c51278861a92ceac0eb6ba58da3/info of 3b2c9c51278861a92ceac0eb6ba58da3 into 6863d90e88074e528626a6a5e511f7dd(size=141.6 K), total size for store is 141.6 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:18:03,201 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:03,201 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., storeName=3b2c9c51278861a92ceac0eb6ba58da3/info, priority=13, startTime=1733411883172; duration=0sec 2024-12-05T15:18:03,202 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:03,202 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 3b2c9c51278861a92ceac0eb6ba58da3:info 2024-12-05T15:18:03,211 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:03,387 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:04,212 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:04,387 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:05,159 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:05,159 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:18:05,164 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/378c9d9070434ed2820cb8ff260d2897 is 1080, key is row0191/info:/1733411883152/Put/seqid=0 2024-12-05T15:18:05,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741867_1043 (size=12516) 2024-12-05T15:18:05,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741867_1043 (size=12516) 2024-12-05T15:18:05,174 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=253 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/378c9d9070434ed2820cb8ff260d2897 2024-12-05T15:18:05,181 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/378c9d9070434ed2820cb8ff260d2897 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/378c9d9070434ed2820cb8ff260d2897 2024-12-05T15:18:05,182 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-05T15:18:05,182 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] ipc.CallRunner(138): callId: 240 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:56170 deadline: 1733411895182, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 2024-12-05T15:18:05,185 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/378c9d9070434ed2820cb8ff260d2897, entries=7, sequenceid=253, filesize=12.2 K 2024-12-05T15:18:05,186 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for 3b2c9c51278861a92ceac0eb6ba58da3 in 27ms, sequenceid=253, compaction requested=false 2024-12-05T15:18:05,186 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:05,212 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:05,388 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:06,213 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:06,388 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:07,213 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:07,389 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:08,214 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:08,389 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:09,214 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:09,390 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:10,215 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:10,390 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:10,777 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2024-12-05T15:18:11,215 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:11,391 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:11,984 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 1588230740, had cached 0 bytes from a total of 14835 2024-12-05T15:18:12,216 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:12,391 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:13,216 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:13,392 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:14,217 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:14,392 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:15,217 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:15,232 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:15,232 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-05T15:18:15,237 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/5949d68c014e4dd7b362bdf3bb3c5eb0 is 1080, key is row0198/info:/1733411885160/Put/seqid=0 2024-12-05T15:18:15,244 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741868_1044 (size=29806) 2024-12-05T15:18:15,245 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741868_1044 (size=29806) 2024-12-05T15:18:15,245 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=279 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/5949d68c014e4dd7b362bdf3bb3c5eb0 2024-12-05T15:18:15,251 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/5949d68c014e4dd7b362bdf3bb3c5eb0 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/5949d68c014e4dd7b362bdf3bb3c5eb0 2024-12-05T15:18:15,255 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/5949d68c014e4dd7b362bdf3bb3c5eb0, entries=23, sequenceid=279, filesize=29.1 K 2024-12-05T15:18:15,256 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=4.20 KB/4304 for 3b2c9c51278861a92ceac0eb6ba58da3 in 24ms, sequenceid=279, compaction requested=true 2024-12-05T15:18:15,256 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:15,256 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 3b2c9c51278861a92ceac0eb6ba58da3:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:18:15,256 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:15,256 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:18:15,257 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 187302 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:18:15,257 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 3b2c9c51278861a92ceac0eb6ba58da3/info is initiating minor compaction (all files) 2024-12-05T15:18:15,257 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 3b2c9c51278861a92ceac0eb6ba58da3/info in TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:15,257 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/6863d90e88074e528626a6a5e511f7dd, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/378c9d9070434ed2820cb8ff260d2897, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/5949d68c014e4dd7b362bdf3bb3c5eb0] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp, totalSize=182.9 K 2024-12-05T15:18:15,258 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 6863d90e88074e528626a6a5e511f7dd, keycount=129, bloomtype=ROW, size=141.6 K, encoding=NONE, compression=NONE, seqNum=242, earliestPutTs=1733411824740 2024-12-05T15:18:15,258 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 378c9d9070434ed2820cb8ff260d2897, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=253, earliestPutTs=1733411883152 2024-12-05T15:18:15,258 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 5949d68c014e4dd7b362bdf3bb3c5eb0, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=279, earliestPutTs=1733411885160 2024-12-05T15:18:15,270 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 3b2c9c51278861a92ceac0eb6ba58da3#info#compaction#64 average throughput is 54.39 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:18:15,271 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/3671318483004ad4a793de0b4b35ae39 is 1080, key is row0062/info:/1733411824740/Put/seqid=0 2024-12-05T15:18:15,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741869_1045 (size=177452) 2024-12-05T15:18:15,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741869_1045 (size=177452) 2024-12-05T15:18:15,279 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/3671318483004ad4a793de0b4b35ae39 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/3671318483004ad4a793de0b4b35ae39 2024-12-05T15:18:15,284 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 3b2c9c51278861a92ceac0eb6ba58da3/info of 3b2c9c51278861a92ceac0eb6ba58da3 into 3671318483004ad4a793de0b4b35ae39(size=173.3 K), total size for store is 173.3 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:18:15,284 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:15,284 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., storeName=3b2c9c51278861a92ceac0eb6ba58da3/info, priority=13, startTime=1733411895256; duration=0sec 2024-12-05T15:18:15,284 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:15,284 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 3b2c9c51278861a92ceac0eb6ba58da3:info 2024-12-05T15:18:15,393 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:16,218 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:16,393 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:17,218 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:17,241 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:17,241 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=7.36 KB heapSize=8.13 KB 2024-12-05T15:18:17,245 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/01fbbe7eb06c4569b483ea8d09e11295 is 1080, key is row0221/info:/1733411895233/Put/seqid=0 2024-12-05T15:18:17,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741870_1046 (size=12523) 2024-12-05T15:18:17,256 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741870_1046 (size=12523) 2024-12-05T15:18:17,256 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=7.36 KB at sequenceid=290 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/01fbbe7eb06c4569b483ea8d09e11295 2024-12-05T15:18:17,262 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/01fbbe7eb06c4569b483ea8d09e11295 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/01fbbe7eb06c4569b483ea8d09e11295 2024-12-05T15:18:17,265 WARN [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(5069): Region is too busy due to exceeding memstore size limit. org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 at org.apache.hadoop.hbase.regionserver.HRegion.checkResources(HRegion.java:5067) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.lambda$put$9(HRegion.java:3229) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:216) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.HRegion.put(HRegion.java:3222) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.put(RSRpcServices.java:3072) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.RSRpcServices.mutate(RSRpcServices.java:3035) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:43506) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2024-12-05T15:18:17,265 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] ipc.CallRunner(138): callId: 272 service: ClientService methodName: Mutate size: 1.2 K connection: 172.17.0.2:56170 deadline: 1733411907265, exception=org.apache.hadoop.hbase.RegionTooBusyException: Over memstore limit=32.0 K, regionName=3b2c9c51278861a92ceac0eb6ba58da3, server=4ecb0f57b706,43509,1733411800834 2024-12-05T15:18:17,267 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/01fbbe7eb06c4569b483ea8d09e11295, entries=7, sequenceid=290, filesize=12.2 K 2024-12-05T15:18:17,268 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~7.36 KB/7532, heapSize ~8.11 KB/8304, currentSize=23.12 KB/23672 for 3b2c9c51278861a92ceac0eb6ba58da3 in 27ms, sequenceid=290, compaction requested=false 2024-12-05T15:18:17,268 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:17,394 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:18,219 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:18,394 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:18,883 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 817a20dfbf0d1531445f95325bcaf27c, had cached 0 bytes from a total of 23930 2024-12-05T15:18:19,219 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:19,395 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:20,220 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:20,395 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:21,221 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:21,396 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:22,221 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:22,396 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:23,222 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:23,397 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:24,222 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:24,397 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:25,223 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:25,398 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:26,223 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:26,398 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:27,224 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:27,280 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=43509 {}] regionserver.HRegion(8581): Flush requested on 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:27,280 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=24.17 KB heapSize=26.13 KB 2024-12-05T15:18:27,285 DEBUG [MemStoreFlusher.0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/4eb50c5d002840c3b1b88db751ce24ac is 1080, key is row0228/info:/1733411897241/Put/seqid=0 2024-12-05T15:18:27,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741871_1047 (size=29807) 2024-12-05T15:18:27,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741871_1047 (size=29807) 2024-12-05T15:18:27,292 INFO [MemStoreFlusher.0 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.17 KB at sequenceid=316 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/4eb50c5d002840c3b1b88db751ce24ac 2024-12-05T15:18:27,297 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/4eb50c5d002840c3b1b88db751ce24ac as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/4eb50c5d002840c3b1b88db751ce24ac 2024-12-05T15:18:27,302 INFO [MemStoreFlusher.0 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/4eb50c5d002840c3b1b88db751ce24ac, entries=23, sequenceid=316, filesize=29.1 K 2024-12-05T15:18:27,303 INFO [MemStoreFlusher.0 {}] regionserver.HRegion(3040): Finished flush of dataSize ~24.17 KB/24748, heapSize ~26.11 KB/26736, currentSize=6.30 KB/6456 for 3b2c9c51278861a92ceac0eb6ba58da3 in 23ms, sequenceid=316, compaction requested=true 2024-12-05T15:18:27,303 DEBUG [MemStoreFlusher.0 {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:27,303 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(403): Add compact mark for store 3b2c9c51278861a92ceac0eb6ba58da3:info, priority=-2147483648, current under compaction store size is 1 2024-12-05T15:18:27,303 DEBUG [MemStoreFlusher.0 {}] regionserver.CompactSplit(411): Small Compaction requested: system; Because: MemStoreFlusher.0; compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:27,303 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.SortedCompactionPolicy(75): Selecting compaction from 3 store files, 0 compacting, 3 eligible, 16 blocking 2024-12-05T15:18:27,304 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.ExploringCompactionPolicy(116): Exploring compaction algorithm has selected 3 files of size 219782 starting at candidate #0 after considering 1 permutations with 1 in ratio 2024-12-05T15:18:27,304 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1540): 3b2c9c51278861a92ceac0eb6ba58da3/info is initiating minor compaction (all files) 2024-12-05T15:18:27,304 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2351): Starting compaction of 3b2c9c51278861a92ceac0eb6ba58da3/info in TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:27,304 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1176): Starting compaction of [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/3671318483004ad4a793de0b4b35ae39, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/01fbbe7eb06c4569b483ea8d09e11295, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/4eb50c5d002840c3b1b88db751ce24ac] into tmpdir=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp, totalSize=214.6 K 2024-12-05T15:18:27,304 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 3671318483004ad4a793de0b4b35ae39, keycount=159, bloomtype=ROW, size=173.3 K, encoding=NONE, compression=NONE, seqNum=279, earliestPutTs=1733411824740 2024-12-05T15:18:27,305 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 01fbbe7eb06c4569b483ea8d09e11295, keycount=7, bloomtype=ROW, size=12.2 K, encoding=NONE, compression=NONE, seqNum=290, earliestPutTs=1733411895233 2024-12-05T15:18:27,305 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] compactions.Compactor(224): Compacting 4eb50c5d002840c3b1b88db751ce24ac, keycount=23, bloomtype=ROW, size=29.1 K, encoding=NONE, compression=NONE, seqNum=316, earliestPutTs=1733411897241 2024-12-05T15:18:27,317 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] throttle.PressureAwareThroughputController(145): 3b2c9c51278861a92ceac0eb6ba58da3#info#compaction#67 average throughput is 48.49 MB/second, slept 0 time(s) and total slept time is 0 ms. 0 active operations remaining, total limit is 50.00 MB/second 2024-12-05T15:18:27,317 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/472bb138088441c3b758944ff25fb388 is 1080, key is row0062/info:/1733411824740/Put/seqid=0 2024-12-05T15:18:27,320 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741872_1048 (size=210001) 2024-12-05T15:18:27,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741872_1048 (size=210001) 2024-12-05T15:18:27,326 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/472bb138088441c3b758944ff25fb388 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/472bb138088441c3b758944ff25fb388 2024-12-05T15:18:27,331 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HStore(1336): Completed compaction of 3 (all) file(s) in 3b2c9c51278861a92ceac0eb6ba58da3/info of 3b2c9c51278861a92ceac0eb6ba58da3 into 472bb138088441c3b758944ff25fb388(size=205.1 K), total size for store is 205.1 K. This selection was in queue for 0sec, and took 0sec to execute. 2024-12-05T15:18:27,331 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.HRegion(2381): Compaction status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:27,331 INFO [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(669): Completed compaction region=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., storeName=3b2c9c51278861a92ceac0eb6ba58da3/info, priority=13, startTime=1733411907303; duration=0sec 2024-12-05T15:18:27,331 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(701): Status compactionQueue=(longCompactions=0:shortCompactions=0), splitQueue=0 2024-12-05T15:18:27,331 DEBUG [RS:0;4ecb0f57b706:43509-shortCompactions-0 {}] regionserver.CompactSplit$CompactionRunner(719): Remove under compaction mark for store: 3b2c9c51278861a92ceac0eb6ba58da3:info 2024-12-05T15:18:27,399 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:28,224 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:28,399 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:29,225 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:29,287 INFO [Time-limited test {}] wal.AbstractTestLogRolling(285): after writing there are 0 log files 2024-12-05T15:18:29,287 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C43509%2C1733411800834.1733411909287 2024-12-05T15:18:29,294 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411801638 with entries=309, filesize=306.80 KB; new WAL /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411909287 2024-12-05T15:18:29,295 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41935:41935),(127.0.0.1/127.0.0.1:39577:39577)] 2024-12-05T15:18:29,295 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411801638 is not closed yet, will try archiving it next time 2024-12-05T15:18:29,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741833_1009 (size=314174) 2024-12-05T15:18:29,296 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741833_1009 (size=314174) 2024-12-05T15:18:29,298 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 14ad900f35513e0519981bcd843ee526 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-05T15:18:29,313 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526/.tmp/info/26777339705e4ecc8d4c878443233475 is 45, key is default/info:d/1733411802428/Put/seqid=0 2024-12-05T15:18:29,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741874_1050 (size=5037) 2024-12-05T15:18:29,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741874_1050 (size=5037) 2024-12-05T15:18:29,319 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526/.tmp/info/26777339705e4ecc8d4c878443233475 2024-12-05T15:18:29,324 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526/.tmp/info/26777339705e4ecc8d4c878443233475 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526/info/26777339705e4ecc8d4c878443233475 2024-12-05T15:18:29,328 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526/info/26777339705e4ecc8d4c878443233475, entries=2, sequenceid=6, filesize=4.9 K 2024-12-05T15:18:29,329 INFO [Time-limited test {}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 14ad900f35513e0519981bcd843ee526 in 31ms, sequenceid=6, compaction requested=false 2024-12-05T15:18:29,329 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 14ad900f35513e0519981bcd843ee526: 2024-12-05T15:18:29,329 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=2.21 KB heapSize=4.13 KB 2024-12-05T15:18:29,333 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/info/f6b04b917c264fce84b8aedf0eb1b374 is 193, key is TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3./info:regioninfo/1733411827499/Put/seqid=0 2024-12-05T15:18:29,337 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741875_1051 (size=7803) 2024-12-05T15:18:29,337 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741875_1051 (size=7803) 2024-12-05T15:18:29,337 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.21 KB at sequenceid=24 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/info/f6b04b917c264fce84b8aedf0eb1b374 2024-12-05T15:18:29,342 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/.tmp/info/f6b04b917c264fce84b8aedf0eb1b374 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/info/f6b04b917c264fce84b8aedf0eb1b374 2024-12-05T15:18:29,348 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/info/f6b04b917c264fce84b8aedf0eb1b374, entries=16, sequenceid=24, filesize=7.6 K 2024-12-05T15:18:29,349 INFO [Time-limited test {}] regionserver.HRegion(3040): Finished flush of dataSize ~2.21 KB/2260, heapSize ~3.61 KB/3696, currentSize=0 B/0 for 1588230740 in 20ms, sequenceid=24, compaction requested=false 2024-12-05T15:18:29,349 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 1588230740: 2024-12-05T15:18:29,349 INFO [Time-limited test {}] regionserver.HRegion(2837): Flushing 3b2c9c51278861a92ceac0eb6ba58da3 1/1 column families, dataSize=6.30 KB heapSize=7 KB 2024-12-05T15:18:29,353 DEBUG [Time-limited test {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/a0f066120bf94f6db4c9e7ea1929776e is 1080, key is row0251/info:/1733411907281/Put/seqid=0 2024-12-05T15:18:29,361 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741876_1052 (size=11436) 2024-12-05T15:18:29,361 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741876_1052 (size=11436) 2024-12-05T15:18:29,361 INFO [Time-limited test {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=6.30 KB at sequenceid=326 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/a0f066120bf94f6db4c9e7ea1929776e 2024-12-05T15:18:29,367 DEBUG [Time-limited test {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/.tmp/info/a0f066120bf94f6db4c9e7ea1929776e as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/a0f066120bf94f6db4c9e7ea1929776e 2024-12-05T15:18:29,371 INFO [Time-limited test {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/a0f066120bf94f6db4c9e7ea1929776e, entries=6, sequenceid=326, filesize=11.2 K 2024-12-05T15:18:29,372 INFO [Time-limited test {}] regionserver.HRegion(3040): Finished flush of dataSize ~6.30 KB/6456, heapSize ~6.98 KB/7152, currentSize=0 B/0 for 3b2c9c51278861a92ceac0eb6ba58da3 in 23ms, sequenceid=326, compaction requested=false 2024-12-05T15:18:29,372 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:29,372 DEBUG [Time-limited test {}] regionserver.HRegion(2538): Flush status journal for ff80d05b17e1b6958441720b0213cd29: 2024-12-05T15:18:29,372 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C43509%2C1733411800834.1733411909372 2024-12-05T15:18:29,383 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411909287 with entries=4, filesize=1.22 KB; new WAL /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411909372 2024-12-05T15:18:29,383 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:41935:41935),(127.0.0.1/127.0.0.1:39577:39577)] 2024-12-05T15:18:29,383 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411909287 is not closed yet, will try archiving it next time 2024-12-05T15:18:29,383 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411801638 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/oldWALs/4ecb0f57b706%2C43509%2C1733411800834.1733411801638 2024-12-05T15:18:29,384 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [5,000] milli-secs(wait.for.ratio=[1]) 2024-12-05T15:18:29,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741873_1049 (size=1255) 2024-12-05T15:18:29,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741873_1049 (size=1255) 2024-12-05T15:18:29,385 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834/4ecb0f57b706%2C43509%2C1733411800834.1733411909287 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/oldWALs/4ecb0f57b706%2C43509%2C1733411800834.1733411909287 2024-12-05T15:18:29,400 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:29,484 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-05T15:18:29,484 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2024-12-05T15:18:29,484 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7319ca6d to 127.0.0.1:59287 2024-12-05T15:18:29,484 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:29,485 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T15:18:29,485 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=423368454, stopped=false 2024-12-05T15:18:29,485 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=4ecb0f57b706,42659,1733411800791 2024-12-05T15:18:29,486 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:18:29,486 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:18:29,486 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:29,486 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:29,486 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-05T15:18:29,487 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:29,487 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:18:29,487 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:18:29,487 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,43509,1733411800834' ***** 2024-12-05T15:18:29,487 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-05T15:18:29,487 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T15:18:29,488 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(3579): Received CLOSE for 14ad900f35513e0519981bcd843ee526 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(3579): Received CLOSE for 3b2c9c51278861a92ceac0eb6ba58da3 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(3579): Received CLOSE for ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,43509,1733411800834 2024-12-05T15:18:29,488 DEBUG [RS:0;4ecb0f57b706:43509 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:29,488 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 14ad900f35513e0519981bcd843ee526, disabling compactions & flushes 2024-12-05T15:18:29,488 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:18:29,488 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:18:29,488 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. after waiting 0 ms 2024-12-05T15:18:29,488 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T15:18:29,488 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:18:29,489 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1599): Waiting on 4 regions to close 2024-12-05T15:18:29,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1603): Online Regions={14ad900f35513e0519981bcd843ee526=hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526., 1588230740=hbase:meta,,1.1588230740, 3b2c9c51278861a92ceac0eb6ba58da3=TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3., ff80d05b17e1b6958441720b0213cd29=TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29.} 2024-12-05T15:18:29,489 DEBUG [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1629): Waiting on 14ad900f35513e0519981bcd843ee526, 1588230740, 3b2c9c51278861a92ceac0eb6ba58da3, ff80d05b17e1b6958441720b0213cd29 2024-12-05T15:18:29,489 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:18:29,489 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:18:29,489 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:18:29,489 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:18:29,489 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:18:29,493 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/namespace/14ad900f35513e0519981bcd843ee526/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-05T15:18:29,493 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/hbase/meta/1588230740/recovered.edits/27.seqid, newMaxSeqId=27, maxSeqId=1 2024-12-05T15:18:29,494 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 14ad900f35513e0519981bcd843ee526: 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733411802016.14ad900f35513e0519981bcd843ee526. 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 3b2c9c51278861a92ceac0eb6ba58da3, disabling compactions & flushes 2024-12-05T15:18:29,494 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. after waiting 0 ms 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T15:18:29,494 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:18:29,494 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-05T15:18:29,494 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663->hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419-top, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/44bce3b7c8d3499e8962867a05bf579e, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/229055645d304b24865458b85f2d6c17, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/c97b24774b9d442ab6869227aad0989a, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/432c031e67924a878efbec9c8e093035, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/741e3d924335466f9bdbf87436ef2b1c, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/8af7a398f44941eb8c188f298510acbd, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/9490ecd40b03448a86d7d0a1fc54ce81, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/31f7c14f8323441b9691b989810a14bd, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e39056b86c174aa7b20f150676701d79, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/eae7dc227fec41dfb2624feb2d94190c, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e02585c3910048d084020c402d4deb86, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/6863d90e88074e528626a6a5e511f7dd, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/aae3b33e969b4f9bbea8bbbf689f5413, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/378c9d9070434ed2820cb8ff260d2897, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/3671318483004ad4a793de0b4b35ae39, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/5949d68c014e4dd7b362bdf3bb3c5eb0, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/01fbbe7eb06c4569b483ea8d09e11295, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/4eb50c5d002840c3b1b88db751ce24ac] to archive 2024-12-05T15:18:29,495 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-05T15:18:29,496 INFO [regionserver/4ecb0f57b706:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:18:29,497 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663 2024-12-05T15:18:29,498 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/44bce3b7c8d3499e8962867a05bf579e to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/44bce3b7c8d3499e8962867a05bf579e 2024-12-05T15:18:29,500 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/TestLogRolling-testLogRolling=7321143ee598c0e7d9799eff39da3663-45a8e7a068754919a0efbec4b7b18eb0 2024-12-05T15:18:29,501 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/229055645d304b24865458b85f2d6c17 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/229055645d304b24865458b85f2d6c17 2024-12-05T15:18:29,502 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/c97b24774b9d442ab6869227aad0989a to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/c97b24774b9d442ab6869227aad0989a 2024-12-05T15:18:29,503 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/432c031e67924a878efbec9c8e093035 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/432c031e67924a878efbec9c8e093035 2024-12-05T15:18:29,504 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/741e3d924335466f9bdbf87436ef2b1c to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/741e3d924335466f9bdbf87436ef2b1c 2024-12-05T15:18:29,505 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/8af7a398f44941eb8c188f298510acbd to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/8af7a398f44941eb8c188f298510acbd 2024-12-05T15:18:29,506 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/9490ecd40b03448a86d7d0a1fc54ce81 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/9490ecd40b03448a86d7d0a1fc54ce81 2024-12-05T15:18:29,508 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/31f7c14f8323441b9691b989810a14bd to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/31f7c14f8323441b9691b989810a14bd 2024-12-05T15:18:29,509 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e39056b86c174aa7b20f150676701d79 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e39056b86c174aa7b20f150676701d79 2024-12-05T15:18:29,510 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/eae7dc227fec41dfb2624feb2d94190c to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/eae7dc227fec41dfb2624feb2d94190c 2024-12-05T15:18:29,511 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e02585c3910048d084020c402d4deb86 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/e02585c3910048d084020c402d4deb86 2024-12-05T15:18:29,513 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/6863d90e88074e528626a6a5e511f7dd to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/6863d90e88074e528626a6a5e511f7dd 2024-12-05T15:18:29,514 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/aae3b33e969b4f9bbea8bbbf689f5413 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/aae3b33e969b4f9bbea8bbbf689f5413 2024-12-05T15:18:29,515 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/378c9d9070434ed2820cb8ff260d2897 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/378c9d9070434ed2820cb8ff260d2897 2024-12-05T15:18:29,516 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/3671318483004ad4a793de0b4b35ae39 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/3671318483004ad4a793de0b4b35ae39 2024-12-05T15:18:29,517 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/5949d68c014e4dd7b362bdf3bb3c5eb0 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/5949d68c014e4dd7b362bdf3bb3c5eb0 2024-12-05T15:18:29,518 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/01fbbe7eb06c4569b483ea8d09e11295 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/01fbbe7eb06c4569b483ea8d09e11295 2024-12-05T15:18:29,519 DEBUG [StoreCloser-TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/4eb50c5d002840c3b1b88db751ce24ac to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/info/4eb50c5d002840c3b1b88db751ce24ac 2024-12-05T15:18:29,523 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/3b2c9c51278861a92ceac0eb6ba58da3/recovered.edits/329.seqid, newMaxSeqId=329, maxSeqId=112 2024-12-05T15:18:29,524 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:29,524 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 3b2c9c51278861a92ceac0eb6ba58da3: 2024-12-05T15:18:29,524 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRolling,row0062,1733411826814.3b2c9c51278861a92ceac0eb6ba58da3. 2024-12-05T15:18:29,524 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing ff80d05b17e1b6958441720b0213cd29, disabling compactions & flushes 2024-12-05T15:18:29,524 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:18:29,524 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:18:29,524 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. after waiting 0 ms 2024-12-05T15:18:29,524 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:18:29,525 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29.-1 {}] regionserver.HStore(2316): Moving the files [hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663->hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/7321143ee598c0e7d9799eff39da3663/info/b145471a147d41e587987baecf372419-bottom] to archive 2024-12-05T15:18:29,525 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29.-1 {}] backup.HFileArchiver(360): Archiving compacted files. 2024-12-05T15:18:29,526 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-05T15:18:29,526 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-05T15:18:29,527 DEBUG [StoreCloser-TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29.-1 {}] backup.HFileArchiver(596): Archived from FileableStoreFile, hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663 to hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/archive/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/info/b145471a147d41e587987baecf372419.7321143ee598c0e7d9799eff39da3663 2024-12-05T15:18:29,530 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/data/default/TestLogRolling-testLogRolling/ff80d05b17e1b6958441720b0213cd29/recovered.edits/117.seqid, newMaxSeqId=117, maxSeqId=112 2024-12-05T15:18:29,531 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:18:29,531 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for ff80d05b17e1b6958441720b0213cd29: 2024-12-05T15:18:29,531 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed TestLogRolling-testLogRolling,,1733411826814.ff80d05b17e1b6958441720b0213cd29. 2024-12-05T15:18:29,689 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,43509,1733411800834; all regions closed. 2024-12-05T15:18:29,689 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834 2024-12-05T15:18:29,691 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741834_1010 (size=9351) 2024-12-05T15:18:29,691 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741834_1010 (size=9351) 2024-12-05T15:18:29,693 DEBUG [RS:0;4ecb0f57b706:43509 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/oldWALs 2024-12-05T15:18:29,693 INFO [RS:0;4ecb0f57b706:43509 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C43509%2C1733411800834.meta:.meta(num 1733411801978) 2024-12-05T15:18:29,694 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/WALs/4ecb0f57b706,43509,1733411800834 2024-12-05T15:18:29,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741877_1053 (size=1071) 2024-12-05T15:18:29,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741877_1053 (size=1071) 2024-12-05T15:18:29,697 DEBUG [RS:0;4ecb0f57b706:43509 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/oldWALs 2024-12-05T15:18:29,697 INFO [RS:0;4ecb0f57b706:43509 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C43509%2C1733411800834:(num 1733411909372) 2024-12-05T15:18:29,697 DEBUG [RS:0;4ecb0f57b706:43509 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:29,697 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:18:29,698 INFO [RS:0;4ecb0f57b706:43509 {}] hbase.ChoreService(370): Chore service for: regionserver/4ecb0f57b706:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-05T15:18:29,698 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:18:29,698 INFO [RS:0;4ecb0f57b706:43509 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:43509 2024-12-05T15:18:29,700 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/4ecb0f57b706,43509,1733411800834 2024-12-05T15:18:29,700 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:18:29,701 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [4ecb0f57b706,43509,1733411800834] 2024-12-05T15:18:29,701 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 4ecb0f57b706,43509,1733411800834; numProcessing=1 2024-12-05T15:18:29,702 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/4ecb0f57b706,43509,1733411800834 already deleted, retry=false 2024-12-05T15:18:29,702 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 4ecb0f57b706,43509,1733411800834 expired; onlineServers=0 2024-12-05T15:18:29,702 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,42659,1733411800791' ***** 2024-12-05T15:18:29,702 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T15:18:29,702 DEBUG [M:0;4ecb0f57b706:42659 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@791becc3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:18:29,702 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,42659,1733411800791 2024-12-05T15:18:29,702 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,42659,1733411800791; all regions closed. 2024-12-05T15:18:29,702 DEBUG [M:0;4ecb0f57b706:42659 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:29,703 DEBUG [M:0;4ecb0f57b706:42659 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T15:18:29,703 DEBUG [M:0;4ecb0f57b706:42659 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T15:18:29,703 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T15:18:29,703 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411801382 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411801382,5,FailOnTimeoutGroup] 2024-12-05T15:18:29,703 INFO [M:0;4ecb0f57b706:42659 {}] hbase.ChoreService(370): Chore service for: master/4ecb0f57b706:0 had [] on shutdown 2024-12-05T15:18:29,703 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411801382 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411801382,5,FailOnTimeoutGroup] 2024-12-05T15:18:29,703 DEBUG [M:0;4ecb0f57b706:42659 {}] master.HMaster(1733): Stopping service threads 2024-12-05T15:18:29,703 INFO [M:0;4ecb0f57b706:42659 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T15:18:29,703 INFO [M:0;4ecb0f57b706:42659 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T15:18:29,703 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T15:18:29,704 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T15:18:29,704 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:29,704 DEBUG [M:0;4ecb0f57b706:42659 {}] zookeeper.ZKUtil(347): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T15:18:29,704 WARN [M:0;4ecb0f57b706:42659 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T15:18:29,704 INFO [M:0;4ecb0f57b706:42659 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-05T15:18:29,704 INFO [M:0;4ecb0f57b706:42659 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T15:18:29,704 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:18:29,704 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:29,704 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:29,704 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:18:29,704 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:29,704 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:18:29,705 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=66.45 KB heapSize=81.70 KB 2024-12-05T15:18:29,720 DEBUG [M:0;4ecb0f57b706:42659 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/afde92b30f5d455aaadcdad99af17ca4 is 82, key is hbase:meta,,1/info:regioninfo/1733411801996/Put/seqid=0 2024-12-05T15:18:29,725 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741878_1054 (size=5672) 2024-12-05T15:18:29,726 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741878_1054 (size=5672) 2024-12-05T15:18:29,726 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/afde92b30f5d455aaadcdad99af17ca4 2024-12-05T15:18:29,747 DEBUG [M:0;4ecb0f57b706:42659 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed316871bf8e43c39ff406fb86649ae1 is 751, key is \x00\x00\x00\x00\x00\x00\x00\x09/proc:d/1733411802929/Put/seqid=0 2024-12-05T15:18:29,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741879_1055 (size=7286) 2024-12-05T15:18:29,752 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741879_1055 (size=7286) 2024-12-05T15:18:29,752 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65.85 KB at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed316871bf8e43c39ff406fb86649ae1 2024-12-05T15:18:29,757 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for ed316871bf8e43c39ff406fb86649ae1 2024-12-05T15:18:29,778 DEBUG [M:0;4ecb0f57b706:42659 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/2b24b131dca84c22bd550fb2b1e1f9ac is 69, key is 4ecb0f57b706,43509,1733411800834/rs:state/1733411801470/Put/seqid=0 2024-12-05T15:18:29,782 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741880_1056 (size=5156) 2024-12-05T15:18:29,783 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741880_1056 (size=5156) 2024-12-05T15:18:29,783 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/2b24b131dca84c22bd550fb2b1e1f9ac 2024-12-05T15:18:29,801 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:29,801 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:43509-0x10062c0bfb10001, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:29,801 INFO [RS:0;4ecb0f57b706:43509 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,43509,1733411800834; zookeeper connection closed. 2024-12-05T15:18:29,802 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@6a519229 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@6a519229 2024-12-05T15:18:29,802 DEBUG [M:0;4ecb0f57b706:42659 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/977d83d9a5424a968762640a47387938 is 52, key is load_balancer_on/state:d/1733411802562/Put/seqid=0 2024-12-05T15:18:29,802 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-05T15:18:29,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741881_1057 (size=5056) 2024-12-05T15:18:29,807 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741881_1057 (size=5056) 2024-12-05T15:18:29,807 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=164 (bloomFilter=true), to=hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/977d83d9a5424a968762640a47387938 2024-12-05T15:18:29,812 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/afde92b30f5d455aaadcdad99af17ca4 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/afde92b30f5d455aaadcdad99af17ca4 2024-12-05T15:18:29,816 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/afde92b30f5d455aaadcdad99af17ca4, entries=8, sequenceid=164, filesize=5.5 K 2024-12-05T15:18:29,817 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/ed316871bf8e43c39ff406fb86649ae1 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ed316871bf8e43c39ff406fb86649ae1 2024-12-05T15:18:29,820 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for ed316871bf8e43c39ff406fb86649ae1 2024-12-05T15:18:29,820 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/ed316871bf8e43c39ff406fb86649ae1, entries=18, sequenceid=164, filesize=7.1 K 2024-12-05T15:18:29,821 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/2b24b131dca84c22bd550fb2b1e1f9ac as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/2b24b131dca84c22bd550fb2b1e1f9ac 2024-12-05T15:18:29,825 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/2b24b131dca84c22bd550fb2b1e1f9ac, entries=1, sequenceid=164, filesize=5.0 K 2024-12-05T15:18:29,826 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/977d83d9a5424a968762640a47387938 as hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/977d83d9a5424a968762640a47387938 2024-12-05T15:18:29,829 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34657/user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/977d83d9a5424a968762640a47387938, entries=1, sequenceid=164, filesize=4.9 K 2024-12-05T15:18:29,830 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(3040): Finished flush of dataSize ~66.45 KB/68043, heapSize ~81.63 KB/83592, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 126ms, sequenceid=164, compaction requested=false 2024-12-05T15:18:29,833 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:29,833 DEBUG [M:0;4ecb0f57b706:42659 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:18:29,833 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/7b888c2b-bf29-24ec-9692-4ac5d0c43cf2/MasterData/WALs/4ecb0f57b706,42659,1733411800791 2024-12-05T15:18:29,835 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43711 is added to blk_1073741830_1006 (size=79272) 2024-12-05T15:18:29,836 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:39039 is added to blk_1073741830_1006 (size=79272) 2024-12-05T15:18:29,836 INFO [M:0;4ecb0f57b706:42659 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-05T15:18:29,836 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:18:29,836 INFO [M:0;4ecb0f57b706:42659 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:42659 2024-12-05T15:18:29,838 DEBUG [M:0;4ecb0f57b706:42659 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/4ecb0f57b706,42659,1733411800791 already deleted, retry=false 2024-12-05T15:18:29,939 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:29,939 INFO [M:0;4ecb0f57b706:42659 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,42659,1733411800791; zookeeper connection closed. 2024-12-05T15:18:29,939 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:42659-0x10062c0bfb10000, quorum=127.0.0.1:59287, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:29,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@33dd8821{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:18:29,942 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@6e41ff6b{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:18:29,942 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:18:29,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@65cd88a1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:18:29,942 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@325f9d5d{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir/,STOPPED} 2024-12-05T15:18:29,944 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:18:29,944 WARN [BP-1838888116-172.17.0.2-1733411800140 heartbeating to localhost/127.0.0.1:34657 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:18:29,944 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:18:29,944 WARN [BP-1838888116-172.17.0.2-1733411800140 heartbeating to localhost/127.0.0.1:34657 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1838888116-172.17.0.2-1733411800140 (Datanode Uuid 59a69e00-1e13-44f8-81d6-b064a4e379ba) service to localhost/127.0.0.1:34657 2024-12-05T15:18:29,944 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data3/current/BP-1838888116-172.17.0.2-1733411800140 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:29,945 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data4/current/BP-1838888116-172.17.0.2-1733411800140 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:29,945 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:18:29,947 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@555a9157{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:18:29,947 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@86b8092{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:18:29,947 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:18:29,948 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@50bb1e61{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:18:29,948 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@15df8ed8{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir/,STOPPED} 2024-12-05T15:18:29,949 WARN [BP-1838888116-172.17.0.2-1733411800140 heartbeating to localhost/127.0.0.1:34657 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:18:29,949 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:18:29,949 WARN [BP-1838888116-172.17.0.2-1733411800140 heartbeating to localhost/127.0.0.1:34657 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1838888116-172.17.0.2-1733411800140 (Datanode Uuid 903ba1ae-98e8-488f-806e-3e814922ab8b) service to localhost/127.0.0.1:34657 2024-12-05T15:18:29,949 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:18:29,949 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data1/current/BP-1838888116-172.17.0.2-1733411800140 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:29,950 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/cluster_93f903d5-6ae3-d71f-465e-b0d3fb7c5083/dfs/data/data2/current/BP-1838888116-172.17.0.2-1733411800140 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:29,950 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:18:29,957 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@17dd071d{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:18:29,958 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@23ed981c{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:18:29,958 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:18:29,958 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@32cfb9e6{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:18:29,958 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3378fb3{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir/,STOPPED} 2024-12-05T15:18:29,965 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-05T15:18:29,993 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-05T15:18:30,002 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRolling Thread=126 (was 113) - Thread LEAK? -, OpenFileDescriptor=487 (was 467) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=107 (was 130), ProcessCount=11 (was 11), AvailableMemoryMB=8341 (was 8425) 2024-12-05T15:18:30,009 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: regionserver.wal.TestLogRolling#testLogRollOnNothingWritten Thread=127, OpenFileDescriptor=487, MaxFileDescriptor=1048576, SystemLoadAverage=107, ProcessCount=11, AvailableMemoryMB=8341 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=1, rsPorts=, rsClass=null, numDataNodes=2, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.log.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.log.dir so I do NOT create it in target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(451): System.getProperty("hadoop.tmp.dir") already set to: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/6800e97d-9d76-1d8c-7c63-14b9ba9f94d5/hadoop.tmp.dir so I do NOT create it in target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d, deleteOnExit=true 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/test.cache.data in system properties and HBase conf 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.tmp.dir in system properties and HBase conf 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.log.dir in system properties and HBase conf 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/mapreduce.cluster.local.dir in system properties and HBase conf 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/mapreduce.cluster.temp.dir in system properties and HBase conf 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2024-12-05T15:18:30,010 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:18:30,010 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/yarn.nodemanager.log-dirs in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/nfs.dump.dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/java.io.tmpdir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/dfs.journalnode.edits.dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2024-12-05T15:18:30,011 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2024-12-05T15:18:30,024 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:18:30,119 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:18:30,124 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:18:30,125 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:18:30,125 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:18:30,125 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:18:30,126 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:18:30,126 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6fa79fad{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:18:30,127 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@102afeed{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:18:30,225 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:30,286 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5a2cc35e{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/java.io.tmpdir/jetty-localhost-38019-hadoop-hdfs-3_4_1-tests_jar-_-any-7739280054740675637/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:18:30,287 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3e7f45c{HTTP/1.1, (http/1.1)}{localhost:38019} 2024-12-05T15:18:30,287 INFO [Time-limited test {}] server.Server(415): Started @401102ms 2024-12-05T15:18:30,299 WARN [Time-limited test {}] blockmanagement.DatanodeManager(468): The given interval for marking stale datanode = 30000, which is larger than heartbeat expire interval 20000. 2024-12-05T15:18:30,353 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:18:30,356 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:18:30,357 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:18:30,357 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:18:30,357 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:18:30,357 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@72d721a0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:18:30,357 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@7d4d798b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:18:30,400 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:30,471 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@318803da{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/java.io.tmpdir/jetty-localhost-37741-hadoop-hdfs-3_4_1-tests_jar-_-any-11962430009369632485/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:18:30,472 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@20584967{HTTP/1.1, (http/1.1)}{localhost:37741} 2024-12-05T15:18:30,472 INFO [Time-limited test {}] server.Server(415): Started @401288ms 2024-12-05T15:18:30,473 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:18:30,504 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2024-12-05T15:18:30,506 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2024-12-05T15:18:30,507 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2024-12-05T15:18:30,507 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2024-12-05T15:18:30,507 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2024-12-05T15:18:30,508 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4a04c7dc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.log.dir/,AVAILABLE} 2024-12-05T15:18:30,508 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3adf0e31{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2024-12-05T15:18:30,569 WARN [Thread-2208 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data1/current/BP-48232372-172.17.0.2-1733411910048/current, will proceed with Du for space computation calculation, 2024-12-05T15:18:30,569 WARN [Thread-2209 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data2/current/BP-48232372-172.17.0.2-1733411910048/current, will proceed with Du for space computation calculation, 2024-12-05T15:18:30,593 WARN [Thread-2187 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:18:30,595 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9a2c2c3b5dff6dd9 with lease ID 0xa9156834a9434c11: Processing first storage report for DS-db5851db-1462-4006-84e4-dd3d985a79d1 from datanode DatanodeRegistration(127.0.0.1:46113, datanodeUuid=7687825e-8d49-4b8f-97c3-5d0df503ef96, infoPort=46155, infoSecurePort=0, ipcPort=38991, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048) 2024-12-05T15:18:30,596 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9a2c2c3b5dff6dd9 with lease ID 0xa9156834a9434c11: from storage DS-db5851db-1462-4006-84e4-dd3d985a79d1 node DatanodeRegistration(127.0.0.1:46113, datanodeUuid=7687825e-8d49-4b8f-97c3-5d0df503ef96, infoPort=46155, infoSecurePort=0, ipcPort=38991, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:18:30,596 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x9a2c2c3b5dff6dd9 with lease ID 0xa9156834a9434c11: Processing first storage report for DS-40281b29-8c08-4bdb-8a4a-261e0519628d from datanode DatanodeRegistration(127.0.0.1:46113, datanodeUuid=7687825e-8d49-4b8f-97c3-5d0df503ef96, infoPort=46155, infoSecurePort=0, ipcPort=38991, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048) 2024-12-05T15:18:30,596 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x9a2c2c3b5dff6dd9 with lease ID 0xa9156834a9434c11: from storage DS-40281b29-8c08-4bdb-8a4a-261e0519628d node DatanodeRegistration(127.0.0.1:46113, datanodeUuid=7687825e-8d49-4b8f-97c3-5d0df503ef96, infoPort=46155, infoSecurePort=0, ipcPort=38991, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:18:30,630 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@521cb856{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/java.io.tmpdir/jetty-localhost-40707-hadoop-hdfs-3_4_1-tests_jar-_-any-10114044755439984027/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:18:30,630 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@60ec5545{HTTP/1.1, (http/1.1)}{localhost:40707} 2024-12-05T15:18:30,630 INFO [Time-limited test {}] server.Server(415): Started @401446ms 2024-12-05T15:18:30,631 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2024-12-05T15:18:30,724 WARN [Thread-2234 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data3/current/BP-48232372-172.17.0.2-1733411910048/current, will proceed with Du for space computation calculation, 2024-12-05T15:18:30,724 WARN [Thread-2235 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data4/current/BP-48232372-172.17.0.2-1733411910048/current, will proceed with Du for space computation calculation, 2024-12-05T15:18:30,740 WARN [Thread-2223 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2024-12-05T15:18:30,742 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x12f67450e90c882c with lease ID 0xa9156834a9434c12: Processing first storage report for DS-8572a112-249a-4b15-84c4-bd1c9ffe5640 from datanode DatanodeRegistration(127.0.0.1:36999, datanodeUuid=29e8e259-b166-437c-b6b9-c09e806bf318, infoPort=37861, infoSecurePort=0, ipcPort=33565, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048) 2024-12-05T15:18:30,742 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x12f67450e90c882c with lease ID 0xa9156834a9434c12: from storage DS-8572a112-249a-4b15-84c4-bd1c9ffe5640 node DatanodeRegistration(127.0.0.1:36999, datanodeUuid=29e8e259-b166-437c-b6b9-c09e806bf318, infoPort=37861, infoSecurePort=0, ipcPort=33565, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:18:30,742 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x12f67450e90c882c with lease ID 0xa9156834a9434c12: Processing first storage report for DS-9ce18e04-3a6d-4175-a682-e49e83070544 from datanode DatanodeRegistration(127.0.0.1:36999, datanodeUuid=29e8e259-b166-437c-b6b9-c09e806bf318, infoPort=37861, infoSecurePort=0, ipcPort=33565, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048) 2024-12-05T15:18:30,742 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x12f67450e90c882c with lease ID 0xa9156834a9434c12: from storage DS-9ce18e04-3a6d-4175-a682-e49e83070544 node DatanodeRegistration(127.0.0.1:36999, datanodeUuid=29e8e259-b166-437c-b6b9-c09e806bf318, infoPort=37861, infoSecurePort=0, ipcPort=33565, storageInfo=lv=-57;cid=testClusterID;nsid=1391423685;c=1733411910048), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2024-12-05T15:18:30,752 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353 2024-12-05T15:18:30,755 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/zookeeper_0, clientPort=60929, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2024-12-05T15:18:30,756 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=60929 2024-12-05T15:18:30,756 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:30,757 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:30,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:18:30,771 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741825_1001 (size=7) 2024-12-05T15:18:30,771 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca with version=8 2024-12-05T15:18:30,772 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1462): The hbase.fs.tmp.dir is set to hdfs://localhost:35505/user/jenkins/test-data/9fed80c5-2fc1-f732-1c1a-96ee037ae86b/hbase-staging 2024-12-05T15:18:30,773 INFO [Time-limited test {}] client.ConnectionUtils(129): master/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:18:30,773 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:18:30,773 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:18:30,774 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:18:30,774 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:18:30,774 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:18:30,774 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:18:30,774 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:18:30,774 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:39997 2024-12-05T15:18:30,775 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:30,776 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:30,778 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:39997 connecting to ZooKeeper ensemble=127.0.0.1:60929 2024-12-05T15:18:30,786 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:399970x0, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:18:30,786 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:39997-0x10062c26d490000 connected 2024-12-05T15:18:30,810 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:18:30,810 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:18:30,811 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:18:30,815 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39997 2024-12-05T15:18:30,815 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39997 2024-12-05T15:18:30,816 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39997 2024-12-05T15:18:30,818 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39997 2024-12-05T15:18:30,818 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39997 2024-12-05T15:18:30,818 INFO [Time-limited test {}] master.HMaster(488): hbase.rootdir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca, hbase.cluster.distributed=false 2024-12-05T15:18:30,842 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/4ecb0f57b706:0 server-side Connection retries=45 2024-12-05T15:18:30,842 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:18:30,842 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2024-12-05T15:18:30,842 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2024-12-05T15:18:30,842 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2024-12-05T15:18:30,842 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2024-12-05T15:18:30,842 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2024-12-05T15:18:30,842 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2024-12-05T15:18:30,843 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.2:35873 2024-12-05T15:18:30,843 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2024-12-05T15:18:30,844 DEBUG [Time-limited test {}] mob.MobFileCache(124): MobFileCache enabled with cacheSize=1000, evictPeriods=3600sec, evictRemainRatio=0.5 2024-12-05T15:18:30,845 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:30,847 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:30,850 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:35873 connecting to ZooKeeper ensemble=127.0.0.1:60929 2024-12-05T15:18:30,852 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:358730x0, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2024-12-05T15:18:30,853 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:35873-0x10062c26d490001 connected 2024-12-05T15:18:30,853 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:18:30,853 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:18:30,854 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2024-12-05T15:18:30,854 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=35873 2024-12-05T15:18:30,856 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=35873 2024-12-05T15:18:30,856 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=35873 2024-12-05T15:18:30,860 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=35873 2024-12-05T15:18:30,861 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=35873 2024-12-05T15:18:30,862 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2445): Adding backup master ZNode /hbase/backup-masters/4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:30,864 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:18:30,864 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:18:30,864 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:30,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:18:30,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2024-12-05T15:18:30,866 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,867 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:18:30,867 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/4ecb0f57b706,39997,1733411910773 from backup master directory 2024-12-05T15:18:30,868 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2024-12-05T15:18:30,869 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:18:30,869 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:18:30,869 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:30,869 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:30,869 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2024-12-05T15:18:30,874 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;4ecb0f57b706:39997 2024-12-05T15:18:30,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:18:30,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741826_1002 (size=42) 2024-12-05T15:18:30,884 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/hbase.id with ID: 56a3312e-9fc7-44bd-9c2e-5507a481eadb 2024-12-05T15:18:30,892 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:30,895 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,895 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:18:30,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741827_1003 (size=196) 2024-12-05T15:18:30,901 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2024-12-05T15:18:30,902 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2024-12-05T15:18:30,902 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:18:30,910 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:18:30,911 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741828_1004 (size=1189) 2024-12-05T15:18:30,911 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store 2024-12-05T15:18:30,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:18:30,920 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741829_1005 (size=34) 2024-12-05T15:18:30,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:18:30,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:18:30,921 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:30,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:30,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:18:30,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:30,921 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:30,921 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:18:30,922 WARN [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/.initializing 2024-12-05T15:18:30,922 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/WALs/4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:30,924 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C39997%2C1733411910773, suffix=, logDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/WALs/4ecb0f57b706,39997,1733411910773, archiveDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/oldWALs, maxLogs=10 2024-12-05T15:18:30,925 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C39997%2C1733411910773.1733411910924 2024-12-05T15:18:30,932 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/WALs/4ecb0f57b706,39997,1733411910773/4ecb0f57b706%2C39997%2C1733411910773.1733411910924 2024-12-05T15:18:30,932 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46155:46155),(127.0.0.1/127.0.0.1:37861:37861)] 2024-12-05T15:18:30,932 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:18:30,932 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:18:30,932 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,932 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,934 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,935 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2024-12-05T15:18:30,935 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:30,936 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:18:30,936 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,937 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2024-12-05T15:18:30,937 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:30,937 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:18:30,938 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,939 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2024-12-05T15:18:30,939 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:30,939 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:18:30,939 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,940 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2024-12-05T15:18:30,940 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:30,941 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:18:30,941 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,942 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,943 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2024-12-05T15:18:30,944 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2024-12-05T15:18:30,946 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:18:30,947 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=823081, jitterRate=0.04660278558731079}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2024-12-05T15:18:30,947 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:18:30,947 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2024-12-05T15:18:30,950 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3a524156, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:18:30,951 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(882): No meta location available on zookeeper, skip migrating... 2024-12-05T15:18:30,951 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2024-12-05T15:18:30,951 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2024-12-05T15:18:30,951 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2024-12-05T15:18:30,951 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 0 msec 2024-12-05T15:18:30,952 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 0 msec 2024-12-05T15:18:30,952 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2024-12-05T15:18:30,953 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2024-12-05T15:18:30,954 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2024-12-05T15:18:30,956 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2024-12-05T15:18:30,956 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2024-12-05T15:18:30,957 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2024-12-05T15:18:30,958 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2024-12-05T15:18:30,958 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2024-12-05T15:18:30,959 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2024-12-05T15:18:30,960 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2024-12-05T15:18:30,961 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2024-12-05T15:18:30,962 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2024-12-05T15:18:30,963 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2024-12-05T15:18:30,965 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2024-12-05T15:18:30,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:18:30,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2024-12-05T15:18:30,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,966 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,966 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(826): Active/primary master=4ecb0f57b706,39997,1733411910773, sessionid=0x10062c26d490000, setting cluster-up flag (Was=false) 2024-12-05T15:18:30,969 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,969 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,974 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2024-12-05T15:18:30,975 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:30,978 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,978 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:30,983 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2024-12-05T15:18:30,983 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2024-12-05T15:18:30,986 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2024-12-05T15:18:30,986 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: 4ecb0f57b706,39997,1733411910773 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=5, maxPoolSize=5 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/4ecb0f57b706:0, corePoolSize=10, maxPoolSize=10 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:18:30,986 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:30,987 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1733411940987 2024-12-05T15:18:30,987 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2024-12-05T15:18:30,987 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2024-12-05T15:18:30,987 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:18:30,987 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2024-12-05T15:18:30,987 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2024-12-05T15:18:30,987 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2024-12-05T15:18:30,987 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2024-12-05T15:18:30,987 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2024-12-05T15:18:30,988 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:30,988 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:30,988 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:18:30,989 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2024-12-05T15:18:30,989 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2024-12-05T15:18:30,989 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2024-12-05T15:18:30,990 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2024-12-05T15:18:30,990 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2024-12-05T15:18:30,992 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411910990,5,FailOnTimeoutGroup] 2024-12-05T15:18:30,992 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411910992,5,FailOnTimeoutGroup] 2024-12-05T15:18:30,992 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:30,992 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1680): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2024-12-05T15:18:30,992 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:30,992 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:30,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:18:30,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741831_1007 (size=1039) 2024-12-05T15:18:30,996 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2024-12-05T15:18:30,996 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca 2024-12-05T15:18:31,002 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:18:31,002 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741832_1008 (size=32) 2024-12-05T15:18:31,003 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:18:31,004 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:18:31,005 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:18:31,005 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,005 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:18:31,005 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:18:31,006 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:18:31,006 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,006 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:18:31,007 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:18:31,007 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:18:31,007 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,008 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:18:31,008 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740 2024-12-05T15:18:31,009 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740 2024-12-05T15:18:31,010 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:18:31,011 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:18:31,013 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:18:31,013 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=775652, jitterRate=-0.013708218932151794}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:18:31,013 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:18:31,013 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:18:31,013 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:18:31,013 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:18:31,013 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:18:31,013 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:18:31,014 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:18:31,014 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:18:31,015 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2024-12-05T15:18:31,015 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2024-12-05T15:18:31,015 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2024-12-05T15:18:31,016 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2024-12-05T15:18:31,016 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2024-12-05T15:18:31,074 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;4ecb0f57b706:35873 2024-12-05T15:18:31,075 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1008): ClusterId : 56a3312e-9fc7-44bd-9c2e-5507a481eadb 2024-12-05T15:18:31,075 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2024-12-05T15:18:31,077 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2024-12-05T15:18:31,077 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2024-12-05T15:18:31,079 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2024-12-05T15:18:31,079 DEBUG [RS:0;4ecb0f57b706:35873 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3e0e8607, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:18:31,079 DEBUG [RS:0;4ecb0f57b706:35873 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7691c8cf, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:18:31,080 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2024-12-05T15:18:31,080 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2024-12-05T15:18:31,080 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1090): About to register with Master. 2024-12-05T15:18:31,080 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(3073): reportForDuty to master=4ecb0f57b706,39997,1733411910773 with isa=4ecb0f57b706/172.17.0.2:35873, startcode=1733411910841 2024-12-05T15:18:31,080 DEBUG [RS:0;4ecb0f57b706:35873 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2024-12-05T15:18:31,082 INFO [RS-EventLoopGroup-14-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:53427, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.6 (auth:SIMPLE), service=RegionServerStatusService 2024-12-05T15:18:31,082 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39997 {}] master.ServerManager(332): Checking decommissioned status of RegionServer 4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,082 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=39997 {}] master.ServerManager(486): Registering regionserver=4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,084 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca 2024-12-05T15:18:31,084 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:34881 2024-12-05T15:18:31,084 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2024-12-05T15:18:31,085 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:18:31,085 DEBUG [RS:0;4ecb0f57b706:35873 {}] zookeeper.ZKUtil(111): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,085 WARN [RS:0;4ecb0f57b706:35873 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2024-12-05T15:18:31,085 INFO [RS:0;4ecb0f57b706:35873 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:18:31,085 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,086 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [4ecb0f57b706,35873,1733411910841] 2024-12-05T15:18:31,088 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2024-12-05T15:18:31,088 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2024-12-05T15:18:31,090 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2024-12-05T15:18:31,090 INFO [RS:0;4ecb0f57b706:35873 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2024-12-05T15:18:31,090 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,090 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2024-12-05T15:18:31,091 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/4ecb0f57b706:0, corePoolSize=2, maxPoolSize=2 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/4ecb0f57b706:0, corePoolSize=1, maxPoolSize=1 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:18:31,091 DEBUG [RS:0;4ecb0f57b706:35873 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/4ecb0f57b706:0, corePoolSize=3, maxPoolSize=3 2024-12-05T15:18:31,092 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,092 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,092 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,092 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,092 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35873,1733411910841-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:18:31,106 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2024-12-05T15:18:31,106 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,35873,1733411910841-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,120 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.Replication(204): 4ecb0f57b706,35873,1733411910841 started 2024-12-05T15:18:31,120 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1767): Serving as 4ecb0f57b706,35873,1733411910841, RpcServer on 4ecb0f57b706/172.17.0.2:35873, sessionid=0x10062c26d490001 2024-12-05T15:18:31,120 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2024-12-05T15:18:31,120 DEBUG [RS:0;4ecb0f57b706:35873 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager 4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,120 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,35873,1733411910841' 2024-12-05T15:18:31,120 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2024-12-05T15:18:31,121 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2024-12-05T15:18:31,121 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2024-12-05T15:18:31,121 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2024-12-05T15:18:31,121 DEBUG [RS:0;4ecb0f57b706:35873 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager 4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,121 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member '4ecb0f57b706,35873,1733411910841' 2024-12-05T15:18:31,121 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2024-12-05T15:18:31,121 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2024-12-05T15:18:31,122 DEBUG [RS:0;4ecb0f57b706:35873 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2024-12-05T15:18:31,122 INFO [RS:0;4ecb0f57b706:35873 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2024-12-05T15:18:31,122 INFO [RS:0;4ecb0f57b706:35873 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2024-12-05T15:18:31,167 WARN [4ecb0f57b706:39997 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2024-12-05T15:18:31,224 INFO [RS:0;4ecb0f57b706:35873 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C35873%2C1733411910841, suffix=, logDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/4ecb0f57b706,35873,1733411910841, archiveDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/oldWALs, maxLogs=32 2024-12-05T15:18:31,224 INFO [RS:0;4ecb0f57b706:35873 {}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C35873%2C1733411910841.1733411911224 2024-12-05T15:18:31,226 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:31,230 INFO [RS:0;4ecb0f57b706:35873 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/4ecb0f57b706,35873,1733411910841/4ecb0f57b706%2C35873%2C1733411910841.1733411911224 2024-12-05T15:18:31,230 DEBUG [RS:0;4ecb0f57b706:35873 {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46155:46155),(127.0.0.1/127.0.0.1:37861:37861)] 2024-12-05T15:18:31,401 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:31,417 DEBUG [4ecb0f57b706:39997 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=1, allServersCount=1 2024-12-05T15:18:31,417 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,418 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,35873,1733411910841, state=OPENING 2024-12-05T15:18:31,420 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2024-12-05T15:18:31,421 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:31,421 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:31,422 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:18:31,422 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:18:31,422 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=4ecb0f57b706,35873,1733411910841}] 2024-12-05T15:18:31,574 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,574 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2024-12-05T15:18:31,576 INFO [RS-EventLoopGroup-15-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60874, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2024-12-05T15:18:31,579 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2024-12-05T15:18:31,579 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:18:31,581 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=4ecb0f57b706%2C35873%2C1733411910841.meta, suffix=.meta, logDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/4ecb0f57b706,35873,1733411910841, archiveDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/oldWALs, maxLogs=32 2024-12-05T15:18:31,581 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor 4ecb0f57b706%2C35873%2C1733411910841.meta.1733411911581.meta 2024-12-05T15:18:31,592 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/4ecb0f57b706,35873,1733411910841/4ecb0f57b706%2C35873%2C1733411910841.meta.1733411911581.meta 2024-12-05T15:18:31,592 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46155:46155),(127.0.0.1/127.0.0.1:37861:37861)] 2024-12-05T15:18:31,592 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:18:31,592 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2024-12-05T15:18:31,592 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2024-12-05T15:18:31,592 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2024-12-05T15:18:31,592 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2024-12-05T15:18:31,593 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:18:31,593 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2024-12-05T15:18:31,593 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2024-12-05T15:18:31,595 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2024-12-05T15:18:31,595 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2024-12-05T15:18:31,595 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,596 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:18:31,596 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2024-12-05T15:18:31,596 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2024-12-05T15:18:31,597 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,597 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:18:31,597 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2024-12-05T15:18:31,597 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2024-12-05T15:18:31,598 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,598 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2024-12-05T15:18:31,599 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740 2024-12-05T15:18:31,600 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740 2024-12-05T15:18:31,601 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (16.0 M)) instead. 2024-12-05T15:18:31,602 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2024-12-05T15:18:31,603 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=857651, jitterRate=0.09056028723716736}}}, FlushLargeStoresPolicy{flushSizeLowerBound=16777216} 2024-12-05T15:18:31,603 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2024-12-05T15:18:31,603 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1733411911574 2024-12-05T15:18:31,605 DEBUG [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2024-12-05T15:18:31,605 INFO [RS_OPEN_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2024-12-05T15:18:31,606 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,606 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as 4ecb0f57b706,35873,1733411910841, state=OPEN 2024-12-05T15:18:31,611 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:18:31,611 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2024-12-05T15:18:31,611 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:18:31,611 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2024-12-05T15:18:31,613 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2024-12-05T15:18:31,613 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=4ecb0f57b706,35873,1733411910841 in 189 msec 2024-12-05T15:18:31,614 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2024-12-05T15:18:31,614 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 598 msec 2024-12-05T15:18:31,616 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 630 msec 2024-12-05T15:18:31,616 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1088): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1733411911616, completionTime=-1 2024-12-05T15:18:31,616 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=1; waited=0ms, expected min=1 server(s), max=1 server(s), master is running 2024-12-05T15:18:31,616 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2024-12-05T15:18:31,617 DEBUG [hconnection-0x25d6e231-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:18:31,618 INFO [RS-EventLoopGroup-15-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60880, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:18:31,619 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=1 2024-12-05T15:18:31,619 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1733411971619 2024-12-05T15:18:31,619 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1733412031619 2024-12-05T15:18:31,619 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 2 msec 2024-12-05T15:18:31,624 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39997,1733411910773-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,624 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39997,1733411910773-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,624 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39997,1733411910773-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,624 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-4ecb0f57b706:39997, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,624 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:31,624 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2024-12-05T15:18:31,624 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(2425): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2024-12-05T15:18:31,625 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2024-12-05T15:18:31,625 DEBUG [master/4ecb0f57b706:0.Chore.1 {}] janitor.CatalogJanitor(179): 2024-12-05T15:18:31,626 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2024-12-05T15:18:31,626 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,626 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2024-12-05T15:18:31,633 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:18:31,633 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741835_1011 (size=358) 2024-12-05T15:18:31,635 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 040376fe86f6ddd4c0d1b0030f915cbc, NAME => 'hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca 2024-12-05T15:18:31,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:18:31,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741836_1012 (size=42) 2024-12-05T15:18:31,642 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:18:31,642 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 040376fe86f6ddd4c0d1b0030f915cbc, disabling compactions & flushes 2024-12-05T15:18:31,642 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:31,642 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:31,642 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. after waiting 0 ms 2024-12-05T15:18:31,642 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:31,642 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:31,642 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 040376fe86f6ddd4c0d1b0030f915cbc: 2024-12-05T15:18:31,643 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2024-12-05T15:18:31,643 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1733411911643"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1733411911643"}]},"ts":"1733411911643"} 2024-12-05T15:18:31,645 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2024-12-05T15:18:31,645 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2024-12-05T15:18:31,645 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411911645"}]},"ts":"1733411911645"} 2024-12-05T15:18:31,647 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2024-12-05T15:18:31,651 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=040376fe86f6ddd4c0d1b0030f915cbc, ASSIGN}] 2024-12-05T15:18:31,652 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=040376fe86f6ddd4c0d1b0030f915cbc, ASSIGN 2024-12-05T15:18:31,652 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=040376fe86f6ddd4c0d1b0030f915cbc, ASSIGN; state=OFFLINE, location=4ecb0f57b706,35873,1733411910841; forceNewPlan=false, retain=false 2024-12-05T15:18:31,803 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=040376fe86f6ddd4c0d1b0030f915cbc, regionState=OPENING, regionLocation=4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,805 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 040376fe86f6ddd4c0d1b0030f915cbc, server=4ecb0f57b706,35873,1733411910841}] 2024-12-05T15:18:31,957 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to 4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,960 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:31,960 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 040376fe86f6ddd4c0d1b0030f915cbc, NAME => 'hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc.', STARTKEY => '', ENDKEY => ''} 2024-12-05T15:18:31,961 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:31,961 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2024-12-05T15:18:31,961 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:31,961 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:31,962 INFO [StoreOpener-040376fe86f6ddd4c0d1b0030f915cbc-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:31,963 INFO [StoreOpener-040376fe86f6ddd4c0d1b0030f915cbc-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 040376fe86f6ddd4c0d1b0030f915cbc columnFamilyName info 2024-12-05T15:18:31,963 DEBUG [StoreOpener-040376fe86f6ddd4c0d1b0030f915cbc-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2024-12-05T15:18:31,963 INFO [StoreOpener-040376fe86f6ddd4c0d1b0030f915cbc-1 {}] regionserver.HStore(327): Store=040376fe86f6ddd4c0d1b0030f915cbc/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2024-12-05T15:18:31,964 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:31,964 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:31,966 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:31,968 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2024-12-05T15:18:31,968 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 040376fe86f6ddd4c0d1b0030f915cbc; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=16384, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=707251, jitterRate=-0.10068459808826447}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2024-12-05T15:18:31,968 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 040376fe86f6ddd4c0d1b0030f915cbc: 2024-12-05T15:18:31,969 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc., pid=6, masterSystemTime=1733411911957 2024-12-05T15:18:31,970 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:31,970 INFO [RS_OPEN_PRIORITY_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:31,971 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=040376fe86f6ddd4c0d1b0030f915cbc, regionState=OPEN, openSeqNum=2, regionLocation=4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:31,974 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2024-12-05T15:18:31,974 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 040376fe86f6ddd4c0d1b0030f915cbc, server=4ecb0f57b706,35873,1733411910841 in 167 msec 2024-12-05T15:18:31,975 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2024-12-05T15:18:31,975 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=040376fe86f6ddd4c0d1b0030f915cbc, ASSIGN in 323 msec 2024-12-05T15:18:31,976 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2024-12-05T15:18:31,976 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1733411911976"}]},"ts":"1733411911976"} 2024-12-05T15:18:31,977 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2024-12-05T15:18:31,980 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2024-12-05T15:18:31,981 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 356 msec 2024-12-05T15:18:32,026 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2024-12-05T15:18:32,028 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:32,028 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:18:32,028 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:32,032 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2024-12-05T15:18:32,039 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:18:32,042 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 10 msec 2024-12-05T15:18:32,043 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2024-12-05T15:18:32,049 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2024-12-05T15:18:32,052 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 8 msec 2024-12-05T15:18:32,058 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2024-12-05T15:18:32,061 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1218): Master has completed initialization 1.192sec 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39997,1733411910773-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2024-12-05T15:18:32,061 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39997,1733411910773-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2024-12-05T15:18:32,063 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster {}] master.HMaster(1321): Balancer post startup initialization complete, took 0 seconds 2024-12-05T15:18:32,063 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2024-12-05T15:18:32,063 INFO [master/4ecb0f57b706:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=4ecb0f57b706,39997,1733411910773-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2024-12-05T15:18:32,064 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x57d7200f to 127.0.0.1:60929 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@1035f033 2024-12-05T15:18:32,067 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@43f52c0a, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2024-12-05T15:18:32,069 DEBUG [hconnection-0x767af09c-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2024-12-05T15:18:32,070 INFO [RS-EventLoopGroup-15-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.2:60886, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2024-12-05T15:18:32,071 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:32,072 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2024-12-05T15:18:32,074 INFO [Time-limited test {}] master.MasterRpcServices(506): Client=null/null set balanceSwitch=false 2024-12-05T15:18:32,074 INFO [Time-limited test {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.FSHLogProvider 2024-12-05T15:18:32,076 INFO [Time-limited test {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=test.com%2C8080%2C1, suffix=, logDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/test.com,8080,1, archiveDir=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/oldWALs, maxLogs=32 2024-12-05T15:18:32,077 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor test.com%2C8080%2C1.1733411912076 2024-12-05T15:18:32,082 INFO [Time-limited test {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/test.com,8080,1/test.com%2C8080%2C1.1733411912076 2024-12-05T15:18:32,082 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:46155:46155),(127.0.0.1/127.0.0.1:37861:37861)] 2024-12-05T15:18:32,082 INFO [Time-limited test {}] monitor.StreamSlowMonitor(122): New stream slow monitor test.com%2C8080%2C1.1733411912082 2024-12-05T15:18:32,087 INFO [Time-limited test {}] wal.AbstractFSWAL(837): Rolled WAL /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/test.com,8080,1/test.com%2C8080%2C1.1733411912076 with entries=0, filesize=85 B; new WAL /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/test.com,8080,1/test.com%2C8080%2C1.1733411912082 2024-12-05T15:18:32,088 DEBUG [Time-limited test {}] wal.AbstractFSWAL(925): Create new FSHLog writer with pipeline: [(127.0.0.1/127.0.0.1:37861:37861),(127.0.0.1/127.0.0.1:46155:46155)] 2024-12-05T15:18:32,088 DEBUG [Time-limited test {}] wal.AbstractFSWAL(751): hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/test.com,8080,1/test.com%2C8080%2C1.1733411912076 is not closed yet, will try archiving it next time 2024-12-05T15:18:32,088 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/test.com,8080,1 2024-12-05T15:18:32,089 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741837_1013 (size=93) 2024-12-05T15:18:32,090 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741838_1014 (size=93) 2024-12-05T15:18:32,090 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741837_1013 (size=93) 2024-12-05T15:18:32,090 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741838_1014 (size=93) 2024-12-05T15:18:32,093 INFO [WAL-Archive-0 {}] wal.AbstractFSWAL(818): Archiving hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/test.com,8080,1/test.com%2C8080%2C1.1733411912076 to hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/oldWALs/test.com%2C8080%2C1.1733411912076 2024-12-05T15:18:32,096 DEBUG [Time-limited test {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/oldWALs 2024-12-05T15:18:32,096 INFO [Time-limited test {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog test.com%2C8080%2C1:(num 1733411912082) 2024-12-05T15:18:32,096 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2024-12-05T15:18:32,096 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x57d7200f to 127.0.0.1:60929 2024-12-05T15:18:32,096 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:32,096 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2024-12-05T15:18:32,096 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=7985492, stopped=false 2024-12-05T15:18:32,096 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:32,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:18:32,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2024-12-05T15:18:32,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:32,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:32,099 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2024-12-05T15:18:32,099 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:18:32,099 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2024-12-05T15:18:32,100 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:32,100 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,35873,1733411910841' ***** 2024-12-05T15:18:32,100 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2024-12-05T15:18:32,100 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HeapMemoryManager(220): Stopping 2024-12-05T15:18:32,100 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2024-12-05T15:18:32,100 INFO [RS:0;4ecb0f57b706:35873 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2024-12-05T15:18:32,100 INFO [RS:0;4ecb0f57b706:35873 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2024-12-05T15:18:32,100 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(3579): Received CLOSE for 040376fe86f6ddd4c0d1b0030f915cbc 2024-12-05T15:18:32,100 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:32,100 DEBUG [RS:0;4ecb0f57b706:35873 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:32,100 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2024-12-05T15:18:32,101 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2024-12-05T15:18:32,101 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 040376fe86f6ddd4c0d1b0030f915cbc, disabling compactions & flushes 2024-12-05T15:18:32,101 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:32,101 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. after waiting 0 ms 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:32,101 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 040376fe86f6ddd4c0d1b0030f915cbc 1/1 column families, dataSize=78 B heapSize=488 B 2024-12-05T15:18:32,101 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1599): Waiting on 2 regions to close 2024-12-05T15:18:32,101 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1603): Online Regions={1588230740=hbase:meta,,1.1588230740, 040376fe86f6ddd4c0d1b0030f915cbc=hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc.} 2024-12-05T15:18:32,101 DEBUG [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1629): Waiting on 040376fe86f6ddd4c0d1b0030f915cbc, 1588230740 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2024-12-05T15:18:32,101 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2024-12-05T15:18:32,101 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2024-12-05T15:18:32,101 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=1.23 KB heapSize=2.87 KB 2024-12-05T15:18:32,116 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc/.tmp/info/f11359f76c7f491e9eee8729a0bddf6e is 45, key is default/info:d/1733411912036/Put/seqid=0 2024-12-05T15:18:32,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741839_1015 (size=5037) 2024-12-05T15:18:32,121 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741839_1015 (size=5037) 2024-12-05T15:18:32,121 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc/.tmp/info/f11359f76c7f491e9eee8729a0bddf6e 2024-12-05T15:18:32,122 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/.tmp/info/885dfe1e07f840db837c84c829947147 is 143, key is hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc./info:regioninfo/1733411911971/Put/seqid=0 2024-12-05T15:18:32,126 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc/.tmp/info/f11359f76c7f491e9eee8729a0bddf6e as hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc/info/f11359f76c7f491e9eee8729a0bddf6e 2024-12-05T15:18:32,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741840_1016 (size=6595) 2024-12-05T15:18:32,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741840_1016 (size=6595) 2024-12-05T15:18:32,128 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.14 KB at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/.tmp/info/885dfe1e07f840db837c84c829947147 2024-12-05T15:18:32,132 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc/info/f11359f76c7f491e9eee8729a0bddf6e, entries=2, sequenceid=6, filesize=4.9 K 2024-12-05T15:18:32,133 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 040376fe86f6ddd4c0d1b0030f915cbc in 32ms, sequenceid=6, compaction requested=false 2024-12-05T15:18:32,137 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/namespace/040376fe86f6ddd4c0d1b0030f915cbc/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2024-12-05T15:18:32,137 INFO [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:32,137 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 040376fe86f6ddd4c0d1b0030f915cbc: 2024-12-05T15:18:32,137 DEBUG [RS_CLOSE_REGION-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1733411911624.040376fe86f6ddd4c0d1b0030f915cbc. 2024-12-05T15:18:32,154 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/.tmp/table/aa16b864094242f98523afd8228ac5c4 is 51, key is hbase:namespace/table:state/1733411911976/Put/seqid=0 2024-12-05T15:18:32,159 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741841_1017 (size=5242) 2024-12-05T15:18:32,159 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741841_1017 (size=5242) 2024-12-05T15:18:32,159 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=94 B at sequenceid=9 (bloomFilter=true), to=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/.tmp/table/aa16b864094242f98523afd8228ac5c4 2024-12-05T15:18:32,164 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/.tmp/info/885dfe1e07f840db837c84c829947147 as hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/info/885dfe1e07f840db837c84c829947147 2024-12-05T15:18:32,168 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: CompactionChecker was stopped 2024-12-05T15:18:32,169 INFO [regionserver/4ecb0f57b706:0.Chore.1 {}] hbase.ScheduledChore(149): Chore: MemstoreFlusherChore was stopped 2024-12-05T15:18:32,169 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/info/885dfe1e07f840db837c84c829947147, entries=10, sequenceid=9, filesize=6.4 K 2024-12-05T15:18:32,169 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/.tmp/table/aa16b864094242f98523afd8228ac5c4 as hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/table/aa16b864094242f98523afd8228ac5c4 2024-12-05T15:18:32,173 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/table/aa16b864094242f98523afd8228ac5c4, entries=2, sequenceid=9, filesize=5.1 K 2024-12-05T15:18:32,174 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~1.23 KB/1264, heapSize ~2.59 KB/2648, currentSize=0 B/0 for 1588230740 in 73ms, sequenceid=9, compaction requested=false 2024-12-05T15:18:32,178 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/data/hbase/meta/1588230740/recovered.edits/12.seqid, newMaxSeqId=12, maxSeqId=1 2024-12-05T15:18:32,178 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2024-12-05T15:18:32,178 INFO [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2024-12-05T15:18:32,178 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2024-12-05T15:18:32,178 DEBUG [RS_CLOSE_META-regionserver/4ecb0f57b706:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2024-12-05T15:18:32,226 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/MasterData/WALs/4ecb0f57b706,38149,1733411626804/4ecb0f57b706%2C38149%2C1733411626804.1733411626978 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:32,301 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,35873,1733411910841; all regions closed. 2024-12-05T15:18:32,301 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:32,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741834_1010 (size=2484) 2024-12-05T15:18:32,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741834_1010 (size=2484) 2024-12-05T15:18:32,306 DEBUG [RS:0;4ecb0f57b706:35873 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/oldWALs 2024-12-05T15:18:32,306 INFO [RS:0;4ecb0f57b706:35873 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C35873%2C1733411910841.meta:.meta(num 1733411911581) 2024-12-05T15:18:32,306 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/WALs/4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:32,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741833_1009 (size=1414) 2024-12-05T15:18:32,308 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741833_1009 (size=1414) 2024-12-05T15:18:32,310 DEBUG [RS:0;4ecb0f57b706:35873 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/oldWALs 2024-12-05T15:18:32,310 INFO [RS:0;4ecb0f57b706:35873 {}] wal.AbstractFSWAL(1074): Closed WAL: FSHLog 4ecb0f57b706%2C35873%2C1733411910841:(num 1733411911224) 2024-12-05T15:18:32,310 DEBUG [RS:0;4ecb0f57b706:35873 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:32,310 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.LeaseManager(133): Closed leases 2024-12-05T15:18:32,310 INFO [RS:0;4ecb0f57b706:35873 {}] hbase.ChoreService(370): Chore service for: regionserver/4ecb0f57b706:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2024-12-05T15:18:32,310 INFO [regionserver/4ecb0f57b706:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:18:32,311 INFO [RS:0;4ecb0f57b706:35873 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:35873 2024-12-05T15:18:32,313 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/4ecb0f57b706,35873,1733411910841 2024-12-05T15:18:32,313 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2024-12-05T15:18:32,315 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [4ecb0f57b706,35873,1733411910841] 2024-12-05T15:18:32,315 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing 4ecb0f57b706,35873,1733411910841; numProcessing=1 2024-12-05T15:18:32,316 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/4ecb0f57b706,35873,1733411910841 already deleted, retry=false 2024-12-05T15:18:32,316 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; 4ecb0f57b706,35873,1733411910841 expired; onlineServers=0 2024-12-05T15:18:32,316 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server '4ecb0f57b706,39997,1733411910773' ***** 2024-12-05T15:18:32,316 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2024-12-05T15:18:32,317 DEBUG [M:0;4ecb0f57b706:39997 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@50b49435, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=4ecb0f57b706/172.17.0.2:0 2024-12-05T15:18:32,317 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HRegionServer(1224): stopping server 4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:32,317 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HRegionServer(1250): stopping server 4ecb0f57b706,39997,1733411910773; all regions closed. 2024-12-05T15:18:32,317 DEBUG [M:0;4ecb0f57b706:39997 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2024-12-05T15:18:32,317 DEBUG [M:0;4ecb0f57b706:39997 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2024-12-05T15:18:32,317 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2024-12-05T15:18:32,317 DEBUG [M:0;4ecb0f57b706:39997 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2024-12-05T15:18:32,317 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411910992 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.small.0-1733411910992,5,FailOnTimeoutGroup] 2024-12-05T15:18:32,317 DEBUG [master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411910990 {}] cleaner.HFileCleaner(306): Exit Thread[master/4ecb0f57b706:0:becomeActiveMaster-HFileCleaner.large.0-1733411910990,5,FailOnTimeoutGroup] 2024-12-05T15:18:32,317 INFO [M:0;4ecb0f57b706:39997 {}] hbase.ChoreService(370): Chore service for: master/4ecb0f57b706:0 had [] on shutdown 2024-12-05T15:18:32,317 DEBUG [M:0;4ecb0f57b706:39997 {}] master.HMaster(1733): Stopping service threads 2024-12-05T15:18:32,317 INFO [M:0;4ecb0f57b706:39997 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2024-12-05T15:18:32,317 INFO [M:0;4ecb0f57b706:39997 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2024-12-05T15:18:32,317 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2024-12-05T15:18:32,318 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2024-12-05T15:18:32,318 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2024-12-05T15:18:32,318 DEBUG [M:0;4ecb0f57b706:39997 {}] zookeeper.ZKUtil(347): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2024-12-05T15:18:32,318 WARN [M:0;4ecb0f57b706:39997 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2024-12-05T15:18:32,318 INFO [M:0;4ecb0f57b706:39997 {}] assignment.AssignmentManager(391): Stopping assignment manager 2024-12-05T15:18:32,318 INFO [M:0;4ecb0f57b706:39997 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2024-12-05T15:18:32,318 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2024-12-05T15:18:32,318 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2024-12-05T15:18:32,318 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:32,318 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:32,318 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2024-12-05T15:18:32,318 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:32,319 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=25.32 KB heapSize=32.31 KB 2024-12-05T15:18:32,334 DEBUG [M:0;4ecb0f57b706:39997 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/472e974533f248c0b78da421026c3ee6 is 82, key is hbase:meta,,1/info:regioninfo/1733411911606/Put/seqid=0 2024-12-05T15:18:32,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741842_1018 (size=5672) 2024-12-05T15:18:32,339 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741842_1018 (size=5672) 2024-12-05T15:18:32,339 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=504 B at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/472e974533f248c0b78da421026c3ee6 2024-12-05T15:18:32,357 DEBUG [M:0;4ecb0f57b706:39997 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/887d4bb183ed4f5084b67eeaae85076b is 696, key is \x00\x00\x00\x00\x00\x00\x00\x04/proc:d/1733411911981/Put/seqid=0 2024-12-05T15:18:32,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741843_1019 (size=6626) 2024-12-05T15:18:32,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741843_1019 (size=6626) 2024-12-05T15:18:32,362 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24.72 KB at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/887d4bb183ed4f5084b67eeaae85076b 2024-12-05T15:18:32,380 DEBUG [M:0;4ecb0f57b706:39997 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/91157068a2ad42e8b04cff9e38b80865 is 69, key is 4ecb0f57b706,35873,1733411910841/rs:state/1733411911083/Put/seqid=0 2024-12-05T15:18:32,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741844_1020 (size=5156) 2024-12-05T15:18:32,386 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741844_1020 (size=5156) 2024-12-05T15:18:32,386 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=65 B at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/91157068a2ad42e8b04cff9e38b80865 2024-12-05T15:18:32,401 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:32987/user/jenkins/test-data/9fc0ebe2-45be-da29-3d6f-c64a59b1e28b/WALs/4ecb0f57b706,44787,1733411626859/4ecb0f57b706%2C44787%2C1733411626859.meta.1733411627726.meta java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.GeneratedMethodAccessor158.invoke(Unknown Source) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.closeWriter(FSHLog.java:464) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.FSHLog.lambda$doReplaceWriter$0(FSHLog.java:402) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 11 more 2024-12-05T15:18:32,409 DEBUG [M:0;4ecb0f57b706:39997 {}] hfile.HFileWriterImpl(814): Len of the biggest cell in hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/800581f461614915852218ae34b6c083 is 52, key is load_balancer_on/state:d/1733411912073/Put/seqid=0 2024-12-05T15:18:32,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741845_1021 (size=5056) 2024-12-05T15:18:32,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741845_1021 (size=5056) 2024-12-05T15:18:32,415 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=48 B at sequenceid=70 (bloomFilter=true), to=hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/800581f461614915852218ae34b6c083 2024-12-05T15:18:32,415 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:32,415 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:35873-0x10062c26d490001, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:32,415 INFO [RS:0;4ecb0f57b706:35873 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,35873,1733411910841; zookeeper connection closed. 2024-12-05T15:18:32,415 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@471731b3 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@471731b3 2024-12-05T15:18:32,415 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 1 regionserver(s) complete 2024-12-05T15:18:32,419 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/info/472e974533f248c0b78da421026c3ee6 as hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/472e974533f248c0b78da421026c3ee6 2024-12-05T15:18:32,423 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/info/472e974533f248c0b78da421026c3ee6, entries=8, sequenceid=70, filesize=5.5 K 2024-12-05T15:18:32,424 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/proc/887d4bb183ed4f5084b67eeaae85076b as hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/887d4bb183ed4f5084b67eeaae85076b 2024-12-05T15:18:32,427 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/proc/887d4bb183ed4f5084b67eeaae85076b, entries=8, sequenceid=70, filesize=6.5 K 2024-12-05T15:18:32,428 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/rs/91157068a2ad42e8b04cff9e38b80865 as hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/91157068a2ad42e8b04cff9e38b80865 2024-12-05T15:18:32,432 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/rs/91157068a2ad42e8b04cff9e38b80865, entries=1, sequenceid=70, filesize=5.0 K 2024-12-05T15:18:32,433 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/.tmp/state/800581f461614915852218ae34b6c083 as hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/800581f461614915852218ae34b6c083 2024-12-05T15:18:32,436 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:34881/user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/state/800581f461614915852218ae34b6c083, entries=1, sequenceid=70, filesize=4.9 K 2024-12-05T15:18:32,437 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(3040): Finished flush of dataSize ~25.32 KB/25929, heapSize ~32.25 KB/33024, currentSize=0 B/0 for 1595e783b53d99cd5eef43b6debb2682 in 119ms, sequenceid=70, compaction requested=false 2024-12-05T15:18:32,438 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2024-12-05T15:18:32,438 DEBUG [M:0;4ecb0f57b706:39997 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2024-12-05T15:18:32,439 DEBUG [WAL-Shutdown-0 {}] wal.FSHLog(499): Closing WAL writer in /user/jenkins/test-data/411b8aea-2d38-420a-1ead-ebcd1bcb43ca/MasterData/WALs/4ecb0f57b706,39997,1733411910773 2024-12-05T15:18:32,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:36999 is added to blk_1073741830_1006 (size=31030) 2024-12-05T15:18:32,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:46113 is added to blk_1073741830_1006 (size=31030) 2024-12-05T15:18:32,441 INFO [M:0;4ecb0f57b706:39997 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2024-12-05T15:18:32,441 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2024-12-05T15:18:32,441 INFO [M:0;4ecb0f57b706:39997 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.2:39997 2024-12-05T15:18:32,443 DEBUG [M:0;4ecb0f57b706:39997 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/4ecb0f57b706,39997,1733411910773 already deleted, retry=false 2024-12-05T15:18:32,544 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:32,544 INFO [M:0;4ecb0f57b706:39997 {}] regionserver.HRegionServer(1307): Exiting; stopping=4ecb0f57b706,39997,1733411910773; zookeeper connection closed. 2024-12-05T15:18:32,544 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:39997-0x10062c26d490000, quorum=127.0.0.1:60929, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2024-12-05T15:18:32,547 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@521cb856{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:18:32,547 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@60ec5545{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:18:32,547 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:18:32,548 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3adf0e31{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:18:32,548 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4a04c7dc{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.log.dir/,STOPPED} 2024-12-05T15:18:32,549 WARN [BP-48232372-172.17.0.2-1733411910048 heartbeating to localhost/127.0.0.1:34881 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:18:32,549 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:18:32,549 WARN [BP-48232372-172.17.0.2-1733411910048 heartbeating to localhost/127.0.0.1:34881 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-48232372-172.17.0.2-1733411910048 (Datanode Uuid 29e8e259-b166-437c-b6b9-c09e806bf318) service to localhost/127.0.0.1:34881 2024-12-05T15:18:32,549 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:18:32,550 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data3/current/BP-48232372-172.17.0.2-1733411910048 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:32,550 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data4/current/BP-48232372-172.17.0.2-1733411910048 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:32,550 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:18:32,552 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@318803da{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2024-12-05T15:18:32,552 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@20584967{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:18:32,552 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:18:32,552 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@7d4d798b{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:18:32,552 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@72d721a0{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.log.dir/,STOPPED} 2024-12-05T15:18:32,554 WARN [BP-48232372-172.17.0.2-1733411910048 heartbeating to localhost/127.0.0.1:34881 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2024-12-05T15:18:32,554 WARN [BP-48232372-172.17.0.2-1733411910048 heartbeating to localhost/127.0.0.1:34881 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-48232372-172.17.0.2-1733411910048 (Datanode Uuid 7687825e-8d49-4b8f-97c3-5d0df503ef96) service to localhost/127.0.0.1:34881 2024-12-05T15:18:32,554 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2024-12-05T15:18:32,554 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2024-12-05T15:18:32,554 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data1/current/BP-48232372-172.17.0.2-1733411910048 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:32,555 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/cluster_4a2faf37-9b85-42da-fa51-9f92c5fe897d/dfs/data/data2/current/BP-48232372-172.17.0.2-1733411910048 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2024-12-05T15:18:32,555 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2024-12-05T15:18:32,560 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5a2cc35e{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2024-12-05T15:18:32,561 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3e7f45c{HTTP/1.1, (http/1.1)}{localhost:0} 2024-12-05T15:18:32,561 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2024-12-05T15:18:32,561 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@102afeed{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2024-12-05T15:18:32,561 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6fa79fad{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-server/target/test-data/4e1f9472-eb00-a356-cd97-bcda6bcd7353/hadoop.log.dir/,STOPPED} 2024-12-05T15:18:32,567 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2024-12-05T15:18:32,582 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down 2024-12-05T15:18:32,591 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: regionserver.wal.TestLogRolling#testLogRollOnNothingWritten Thread=150 (was 127) - Thread LEAK? -, OpenFileDescriptor=516 (was 487) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=106 (was 107), ProcessCount=11 (was 11), AvailableMemoryMB=8333 (was 8341)