2025-01-08 02:19:59,912 main DEBUG Apache Log4j Core 2.17.2 initializing configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@5e7cd6cc 2025-01-08 02:19:59,928 main DEBUG Took 0.014004 seconds to load 1 plugins from package org.apache.hadoop.hbase.logging 2025-01-08 02:19:59,928 main DEBUG PluginManager 'Core' found 129 plugins 2025-01-08 02:19:59,929 main DEBUG PluginManager 'Level' found 0 plugins 2025-01-08 02:19:59,930 main DEBUG PluginManager 'Lookup' found 16 plugins 2025-01-08 02:19:59,931 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,939 main DEBUG PluginManager 'TypeConverter' found 26 plugins 2025-01-08 02:19:59,955 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.util.MBeans", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,957 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,958 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.logging.TestJul2Slf4j", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,959 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,959 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.zookeeper", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,960 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,961 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSinkAdapter", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,961 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,962 main DEBUG LoggerConfig$Builder(additivity="null", level="ERROR", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsSystemImpl", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,962 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,963 main DEBUG LoggerConfig$Builder(additivity="false", level="WARN", levelAndRefs="null", name="org.apache.directory", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,964 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,964 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.ipc.FailedServers", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,965 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,965 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop.metrics2.impl.MetricsConfig", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,966 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,966 main DEBUG LoggerConfig$Builder(additivity="null", level="INFO", levelAndRefs="null", name="org.apache.hadoop.hbase.ScheduledChore", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,967 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,967 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase.regionserver.RSRpcServices", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,967 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,968 main DEBUG LoggerConfig$Builder(additivity="null", level="WARN", levelAndRefs="null", name="org.apache.hadoop", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,968 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,969 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hadoop.hbase", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,969 main DEBUG Building Plugin[name=logger, class=org.apache.logging.log4j.core.config.LoggerConfig]. 2025-01-08 02:19:59,970 main DEBUG LoggerConfig$Builder(additivity="null", level="DEBUG", levelAndRefs="null", name="org.apache.hbase.thirdparty.io.netty.channel", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,970 main DEBUG Building Plugin[name=root, class=org.apache.logging.log4j.core.config.LoggerConfig$RootLogger]. 2025-01-08 02:19:59,972 main DEBUG LoggerConfig$RootLogger$Builder(additivity="null", level="null", levelAndRefs="INFO,Console", includeLocation="null", ={}, ={}, Configuration(PropertiesConfig), Filter=null) 2025-01-08 02:19:59,974 main DEBUG Building Plugin[name=loggers, class=org.apache.logging.log4j.core.config.LoggersPlugin]. 2025-01-08 02:19:59,976 main DEBUG createLoggers(={org.apache.hadoop.metrics2.util.MBeans, org.apache.hadoop.hbase.logging.TestJul2Slf4j, org.apache.zookeeper, org.apache.hadoop.metrics2.impl.MetricsSinkAdapter, org.apache.hadoop.metrics2.impl.MetricsSystemImpl, org.apache.directory, org.apache.hadoop.hbase.ipc.FailedServers, org.apache.hadoop.metrics2.impl.MetricsConfig, org.apache.hadoop.hbase.ScheduledChore, org.apache.hadoop.hbase.regionserver.RSRpcServices, org.apache.hadoop, org.apache.hadoop.hbase, org.apache.hbase.thirdparty.io.netty.channel, root}) 2025-01-08 02:19:59,977 main DEBUG Building Plugin[name=layout, class=org.apache.logging.log4j.core.layout.PatternLayout]. 2025-01-08 02:19:59,978 main DEBUG PatternLayout$Builder(pattern="%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n", PatternSelector=null, Configuration(PropertiesConfig), Replace=null, charset="null", alwaysWriteExceptions="null", disableAnsi="null", noConsoleNoAnsi="null", header="null", footer="null") 2025-01-08 02:19:59,979 main DEBUG PluginManager 'Converter' found 47 plugins 2025-01-08 02:19:59,989 main DEBUG Building Plugin[name=appender, class=org.apache.hadoop.hbase.logging.HBaseTestAppender]. 2025-01-08 02:19:59,992 main DEBUG HBaseTestAppender$Builder(target="SYSTEM_ERR", maxSize="1G", bufferedIo="null", bufferSize="null", immediateFlush="null", ignoreExceptions="null", PatternLayout(%d{ISO8601} %-5p [%t%notEmpty{ %X}] %C{2}(%L): %m%n), name="Console", Configuration(PropertiesConfig), Filter=null, ={}) 2025-01-08 02:19:59,994 main DEBUG Starting HBaseTestOutputStreamManager SYSTEM_ERR 2025-01-08 02:19:59,994 main DEBUG Building Plugin[name=appenders, class=org.apache.logging.log4j.core.config.AppendersPlugin]. 2025-01-08 02:19:59,995 main DEBUG createAppenders(={Console}) 2025-01-08 02:19:59,996 main DEBUG Configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@5e7cd6cc initialized 2025-01-08 02:19:59,996 main DEBUG Starting configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@5e7cd6cc 2025-01-08 02:19:59,996 main DEBUG Started configuration org.apache.logging.log4j.core.config.properties.PropertiesConfiguration@5e7cd6cc OK. 2025-01-08 02:19:59,997 main DEBUG Shutting down OutputStreamManager SYSTEM_OUT.false.false-1 2025-01-08 02:19:59,998 main DEBUG OutputStream closed 2025-01-08 02:19:59,998 main DEBUG Shut down OutputStreamManager SYSTEM_OUT.false.false-1, all resources released: true 2025-01-08 02:19:59,998 main DEBUG Appender DefaultConsole-1 stopped with status true 2025-01-08 02:19:59,999 main DEBUG Stopped org.apache.logging.log4j.core.config.DefaultConfiguration@434a63ab OK 2025-01-08 02:20:00,075 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6 2025-01-08 02:20:00,077 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=StatusLogger 2025-01-08 02:20:00,078 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=ContextSelector 2025-01-08 02:20:00,080 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name= 2025-01-08 02:20:00,080 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.directory 2025-01-08 02:20:00,081 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSinkAdapter 2025-01-08 02:20:00,081 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.zookeeper 2025-01-08 02:20:00,082 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.logging.TestJul2Slf4j 2025-01-08 02:20:00,082 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsSystemImpl 2025-01-08 02:20:00,083 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.util.MBeans 2025-01-08 02:20:00,083 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase 2025-01-08 02:20:00,084 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop 2025-01-08 02:20:00,084 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ipc.FailedServers 2025-01-08 02:20:00,085 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.regionserver.RSRpcServices 2025-01-08 02:20:00,085 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.metrics2.impl.MetricsConfig 2025-01-08 02:20:00,085 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hbase.thirdparty.io.netty.channel 2025-01-08 02:20:00,086 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Loggers,name=org.apache.hadoop.hbase.ScheduledChore 2025-01-08 02:20:00,087 main DEBUG Registering MBean org.apache.logging.log4j2:type=1dbd16a6,component=Appenders,name=Console 2025-01-08 02:20:00,091 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2025-01-08 02:20:00,091 main DEBUG Reconfiguration complete for context[name=1dbd16a6] at URI jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-logging/target/hbase-logging-2.7.0-SNAPSHOT-tests.jar!/log4j2.properties (org.apache.logging.log4j.core.LoggerContext@4426bff1) with optional ClassLoader: null 2025-01-08 02:20:00,092 main DEBUG Shutdown hook enabled. Registering a new one. 2025-01-08 02:20:00,093 main DEBUG LoggerContext[name=1dbd16a6, org.apache.logging.log4j.core.LoggerContext@4426bff1] started OK. 2025-01-08T02:20:00,112 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestExportSnapshot timeout: 13 mins 2025-01-08 02:20:00,116 main DEBUG AsyncLogger.ThreadNameStrategy=UNCACHED (user specified null, default is UNCACHED) 2025-01-08 02:20:00,116 main DEBUG org.apache.logging.log4j.core.util.SystemClock supports precise timestamps. 2025-01-08T02:20:00,550 DEBUG [main {}] hbase.HBaseTestingUtility(348): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf 2025-01-08T02:20:00,551 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestMobExportSnapshot timeout: 13 mins 2025-01-08T02:20:00,551 INFO [main {}] hbase.HBaseClassTestRule(94): Test class org.apache.hadoop.hbase.snapshot.TestMobSecureExportSnapshot timeout: 13 mins 2025-01-08T02:20:00,584 WARN [Time-limited test {}] util.NativeCodeLoader(60): Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2025-01-08T02:20:00,802 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1126): Starting up minicluster with option: StartMiniClusterOption{numMasters=1, masterClass=null, numRegionServers=3, rsPorts=, rsClass=null, numDataNodes=3, dataNodeHosts=null, numZkServers=1, createRootDir=false, createWALDir=false} 2025-01-08T02:20:00,805 INFO [Time-limited test {}] hbase.HBaseZKTestingUtility(82): Created new mini-cluster data directory: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56, deleteOnExit=true 2025-01-08T02:20:00,805 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1140): STARTING DFS 2025-01-08T02:20:00,807 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/test.cache.data in system properties and HBase conf 2025-01-08T02:20:00,808 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.tmp.dir in system properties and HBase conf 2025-01-08T02:20:00,808 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir in system properties and HBase conf 2025-01-08T02:20:00,809 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/mapreduce.cluster.local.dir in system properties and HBase conf 2025-01-08T02:20:00,809 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/mapreduce.cluster.temp.dir in system properties and HBase conf 2025-01-08T02:20:00,810 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2025-01-08T02:20:00,921 DEBUG [Time-limited test {}] fs.HFileSystem(310): The file system is not a DistributedFileSystem. Skipping on block location reordering 2025-01-08T02:20:00,925 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2025-01-08T02:20:00,926 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2025-01-08T02:20:00,927 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.nodemanager.log-dirs in system properties and HBase conf 2025-01-08T02:20:00,927 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T02:20:00,928 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2025-01-08T02:20:00,928 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2025-01-08T02:20:00,929 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T02:20:00,929 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T02:20:00,930 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2025-01-08T02:20:00,930 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/nfs.dump.dir in system properties and HBase conf 2025-01-08T02:20:00,931 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir in system properties and HBase conf 2025-01-08T02:20:00,931 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T02:20:00,931 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2025-01-08T02:20:00,932 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2025-01-08T02:20:02,152 WARN [Time-limited test {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties 2025-01-08T02:20:02,277 INFO [Time-limited test {}] log.Log(170): Logging initialized @3782ms to org.eclipse.jetty.util.log.Slf4jLog 2025-01-08T02:20:02,431 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:02,676 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:02,792 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:02,792 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:02,797 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T02:20:02,840 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:02,864 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@690f3f23{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:02,865 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6f0daeb1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:03,163 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@603d2b6a{hdfs,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-35271-hadoop-hdfs-3_4_1-tests_jar-_-any-9367895584840344091/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2025-01-08T02:20:03,173 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@3797254c{HTTP/1.1, (http/1.1)}{localhost:35271} 2025-01-08T02:20:03,173 INFO [Time-limited test {}] server.Server(415): Started @4679ms 2025-01-08T02:20:03,948 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:03,962 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:03,971 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:03,972 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:03,974 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T02:20:03,982 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@74ef64a6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:03,990 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@3299e9d3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:04,280 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@5a7079af{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-45059-hadoop-hdfs-3_4_1-tests_jar-_-any-4299646908320161213/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T02:20:04,286 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@76ce3a96{HTTP/1.1, (http/1.1)}{localhost:45059} 2025-01-08T02:20:04,287 INFO [Time-limited test {}] server.Server(415): Started @5793ms 2025-01-08T02:20:04,417 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T02:20:04,637 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:04,649 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:04,697 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:04,697 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:04,698 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T02:20:04,708 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@64a3ece4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:04,709 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@24e59c37{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:04,947 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@6e8b8a6d{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-36645-hadoop-hdfs-3_4_1-tests_jar-_-any-12870627058246723422/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T02:20:04,948 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@62d7732{HTTP/1.1, (http/1.1)}{localhost:36645} 2025-01-08T02:20:04,949 INFO [Time-limited test {}] server.Server(415): Started @6455ms 2025-01-08T02:20:04,951 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T02:20:05,289 WARN [Thread-93 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1/current/BP-1555733941-172.17.0.3-1736302801782/current, will proceed with Du for space computation calculation, 2025-01-08T02:20:05,309 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:05,313 WARN [Thread-96 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2/current/BP-1555733941-172.17.0.3-1736302801782/current, will proceed with Du for space computation calculation, 2025-01-08T02:20:05,332 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:05,346 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:05,346 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:05,346 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T02:20:05,349 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@1887d910{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:05,350 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@178a4d2d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:05,359 WARN [Thread-110 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4/current/BP-1555733941-172.17.0.3-1736302801782/current, will proceed with Du for space computation calculation, 2025-01-08T02:20:05,360 WARN [Thread-109 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3/current/BP-1555733941-172.17.0.3-1736302801782/current, will proceed with Du for space computation calculation, 2025-01-08T02:20:05,415 WARN [Thread-58 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T02:20:05,421 WARN [Thread-83 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T02:20:05,527 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xad2d31074517dc5f with lease ID 0x1d06c9fc3b280c4f: Processing first storage report for DS-c2bc8ca6-8220-462d-8658-7119831cd0d8 from datanode DatanodeRegistration(127.0.0.1:41871, datanodeUuid=66baa9c9-8a2d-42eb-9f59-398f6cb5787c, infoPort=39123, infoSecurePort=0, ipcPort=45047, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782) 2025-01-08T02:20:05,529 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xad2d31074517dc5f with lease ID 0x1d06c9fc3b280c4f: from storage DS-c2bc8ca6-8220-462d-8658-7119831cd0d8 node DatanodeRegistration(127.0.0.1:41871, datanodeUuid=66baa9c9-8a2d-42eb-9f59-398f6cb5787c, infoPort=39123, infoSecurePort=0, ipcPort=45047, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 0, hasStaleStorage: true, processing time: 2 msecs, invalidatedBlocks: 0 2025-01-08T02:20:05,529 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7323392cc96dfb56 with lease ID 0x1d06c9fc3b280c4e: Processing first storage report for DS-ca0c33db-d569-4a43-bb83-bf5f6c1dccb8 from datanode DatanodeRegistration(127.0.0.1:43235, datanodeUuid=d1c88fec-0d29-4195-9b79-b97d3d006240, infoPort=45919, infoSecurePort=0, ipcPort=33591, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782) 2025-01-08T02:20:05,530 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7323392cc96dfb56 with lease ID 0x1d06c9fc3b280c4e: from storage DS-ca0c33db-d569-4a43-bb83-bf5f6c1dccb8 node DatanodeRegistration(127.0.0.1:43235, datanodeUuid=d1c88fec-0d29-4195-9b79-b97d3d006240, infoPort=45919, infoSecurePort=0, ipcPort=33591, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 0, hasStaleStorage: true, processing time: 1 msecs, invalidatedBlocks: 0 2025-01-08T02:20:05,530 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0xad2d31074517dc5f with lease ID 0x1d06c9fc3b280c4f: Processing first storage report for DS-d4cf5599-f280-4d8a-b4b9-34d1b4f6c114 from datanode DatanodeRegistration(127.0.0.1:41871, datanodeUuid=66baa9c9-8a2d-42eb-9f59-398f6cb5787c, infoPort=39123, infoSecurePort=0, ipcPort=45047, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782) 2025-01-08T02:20:05,530 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0xad2d31074517dc5f with lease ID 0x1d06c9fc3b280c4f: from storage DS-d4cf5599-f280-4d8a-b4b9-34d1b4f6c114 node DatanodeRegistration(127.0.0.1:41871, datanodeUuid=66baa9c9-8a2d-42eb-9f59-398f6cb5787c, infoPort=39123, infoSecurePort=0, ipcPort=45047, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T02:20:05,530 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x7323392cc96dfb56 with lease ID 0x1d06c9fc3b280c4e: Processing first storage report for DS-c974eb5c-1097-44f5-b2c5-dba7603fea4b from datanode DatanodeRegistration(127.0.0.1:43235, datanodeUuid=d1c88fec-0d29-4195-9b79-b97d3d006240, infoPort=45919, infoSecurePort=0, ipcPort=33591, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782) 2025-01-08T02:20:05,531 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x7323392cc96dfb56 with lease ID 0x1d06c9fc3b280c4e: from storage DS-c974eb5c-1097-44f5-b2c5-dba7603fea4b node DatanodeRegistration(127.0.0.1:43235, datanodeUuid=d1c88fec-0d29-4195-9b79-b97d3d006240, infoPort=45919, infoSecurePort=0, ipcPort=33591, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T02:20:05,545 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@12bc8767{datanode,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-46313-hadoop-hdfs-3_4_1-tests_jar-_-any-17684396273989816675/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T02:20:05,547 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@e141426{HTTP/1.1, (http/1.1)}{localhost:46313} 2025-01-08T02:20:05,547 INFO [Time-limited test {}] server.Server(415): Started @7053ms 2025-01-08T02:20:05,552 WARN [Time-limited test {}] web.RestCsrfPreventionFilterHandler(75): Got null for restCsrfPreventionFilter - will not do any filtering. 2025-01-08T02:20:05,900 WARN [Thread-139 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5/current/BP-1555733941-172.17.0.3-1736302801782/current, will proceed with Du for space computation calculation, 2025-01-08T02:20:05,902 WARN [Thread-140 {}] impl.BlockPoolSlice(347): dfsUsed file missing in /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6/current/BP-1555733941-172.17.0.3-1736302801782/current, will proceed with Du for space computation calculation, 2025-01-08T02:20:06,041 WARN [Thread-129 {}] datanode.DirectoryScanner(302): dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value above 1000 ms/sec. Assuming default value of -1 2025-01-08T02:20:06,056 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x58e22cc5b42a8f with lease ID 0x1d06c9fc3b280c50: Processing first storage report for DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e from datanode DatanodeRegistration(127.0.0.1:41199, datanodeUuid=5f73ff73-0f7b-4bdf-9032-a33a82e289db, infoPort=41865, infoSecurePort=0, ipcPort=36113, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782) 2025-01-08T02:20:06,056 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x58e22cc5b42a8f with lease ID 0x1d06c9fc3b280c50: from storage DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e node DatanodeRegistration(127.0.0.1:41199, datanodeUuid=5f73ff73-0f7b-4bdf-9032-a33a82e289db, infoPort=41865, infoSecurePort=0, ipcPort=36113, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 0, hasStaleStorage: true, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T02:20:06,057 INFO [Block report processor {}] blockmanagement.BlockManager(2940): BLOCK* processReport 0x58e22cc5b42a8f with lease ID 0x1d06c9fc3b280c50: Processing first storage report for DS-c01f0001-ea9e-4e51-b8a4-76247c985dca from datanode DatanodeRegistration(127.0.0.1:41199, datanodeUuid=5f73ff73-0f7b-4bdf-9032-a33a82e289db, infoPort=41865, infoSecurePort=0, ipcPort=36113, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782) 2025-01-08T02:20:06,057 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x58e22cc5b42a8f with lease ID 0x1d06c9fc3b280c50: from storage DS-c01f0001-ea9e-4e51-b8a4-76247c985dca node DatanodeRegistration(127.0.0.1:41199, datanodeUuid=5f73ff73-0f7b-4bdf-9032-a33a82e289db, infoPort=41865, infoSecurePort=0, ipcPort=36113, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 0, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 0 2025-01-08T02:20:06,389 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(703): Setting hbase.rootdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf 2025-01-08T02:20:06,627 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(259): Started connectionTimeout=30000, dir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/zookeeper_0, clientPort=61593, secureClientPort=-1, dataDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/zookeeper_0/version-2, dataDirSize=457 dataLogDir=/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/zookeeper_0/version-2, dataLogSize=457 tickTime=2000, maxClientCnxns=300, minSessionTimeout=4000, maxSessionTimeout=40000, clientPortListenBacklog=-1, serverId=0 2025-01-08T02:20:06,660 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(284): Started MiniZooKeeperCluster and ran 'stat' on client port=61593 2025-01-08T02:20:06,704 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:06,707 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:07,152 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741825_1001 (size=7) 2025-01-08T02:20:07,153 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741825_1001 (size=7) 2025-01-08T02:20:07,154 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741825_1001 (size=7) 2025-01-08T02:20:07,576 INFO [Time-limited test {}] util.FSUtils(490): Created version file at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 with version=8 2025-01-08T02:20:07,576 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1460): Setting hbase.fs.tmp.dir to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/hbase-staging 2025-01-08T02:20:07,778 DEBUG [Time-limited test {}] channel.MultithreadEventLoopGroup(44): -Dio.netty.eventLoopThreads: 16 2025-01-08T02:20:08,143 INFO [Time-limited test {}] client.ConnectionUtils(129): master/c4e5d52b9759:0 server-side Connection retries=45 2025-01-08T02:20:08,171 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,172 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,172 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T02:20:08,172 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,173 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T02:20:08,360 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.MasterService, hbase.pb.RegionServerStatusService, hbase.pb.LockService, hbase.pb.HbckService, hbase.pb.ClientMetaService, hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T02:20:08,435 INFO [Time-limited test {}] metrics.MetricRegistriesLoader(60): Loaded MetricRegistries class org.apache.hadoop.hbase.metrics.impl.MetricRegistriesImpl 2025-01-08T02:20:08,446 DEBUG [Time-limited test {}] util.ClassSize(228): Using Unsafe to estimate memory layout 2025-01-08T02:20:08,450 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T02:20:08,483 DEBUG [Time-limited test {}] channel.DefaultChannelId(84): -Dio.netty.processId: 18858 (auto-detected) 2025-01-08T02:20:08,484 DEBUG [Time-limited test {}] channel.DefaultChannelId(106): -Dio.netty.machineId: 02:42:ac:ff:fe:11:00:03 (auto-detected) 2025-01-08T02:20:08,516 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:41063 2025-01-08T02:20:08,529 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,533 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,550 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=master:41063 connecting to ZooKeeper ensemble=127.0.0.1:61593 2025-01-08T02:20:08,596 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:410630x0, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T02:20:08,600 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): master:41063-0x1015754588a0000 connected 2025-01-08T02:20:08,654 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T02:20:08,659 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:20:08,677 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T02:20:08,684 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=41063 2025-01-08T02:20:08,685 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=41063 2025-01-08T02:20:08,688 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=41063 2025-01-08T02:20:08,692 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=41063 2025-01-08T02:20:08,692 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=41063 2025-01-08T02:20:08,702 INFO [Time-limited test {}] master.HMaster(499): hbase.rootdir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572, hbase.cluster.distributed=false 2025-01-08T02:20:08,765 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/c4e5d52b9759:0 server-side Connection retries=45 2025-01-08T02:20:08,766 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,766 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,766 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T02:20:08,767 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,767 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T02:20:08,770 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T02:20:08,772 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T02:20:08,773 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:37031 2025-01-08T02:20:08,776 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T02:20:08,781 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T02:20:08,783 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,786 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,792 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:37031 connecting to ZooKeeper ensemble=127.0.0.1:61593 2025-01-08T02:20:08,796 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:370310x0, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T02:20:08,798 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:370310x0, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T02:20:08,798 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:37031-0x1015754588a0001 connected 2025-01-08T02:20:08,800 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:20:08,801 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T02:20:08,805 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=37031 2025-01-08T02:20:08,812 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=37031 2025-01-08T02:20:08,813 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=37031 2025-01-08T02:20:08,813 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=37031 2025-01-08T02:20:08,814 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=37031 2025-01-08T02:20:08,838 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/c4e5d52b9759:0 server-side Connection retries=45 2025-01-08T02:20:08,838 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,838 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,839 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T02:20:08,839 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,839 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T02:20:08,840 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T02:20:08,840 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T02:20:08,843 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:36743 2025-01-08T02:20:08,844 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T02:20:08,854 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T02:20:08,856 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,861 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,866 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:36743 connecting to ZooKeeper ensemble=127.0.0.1:61593 2025-01-08T02:20:08,873 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:367430x0, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T02:20:08,873 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:367430x0, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T02:20:08,875 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:367430x0, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:20:08,876 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:367430x0, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T02:20:08,877 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:36743-0x1015754588a0002 connected 2025-01-08T02:20:08,878 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=36743 2025-01-08T02:20:08,879 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=36743 2025-01-08T02:20:08,879 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=36743 2025-01-08T02:20:08,885 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=36743 2025-01-08T02:20:08,885 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=36743 2025-01-08T02:20:08,914 INFO [Time-limited test {}] client.ConnectionUtils(129): regionserver/c4e5d52b9759:0 server-side Connection retries=45 2025-01-08T02:20:08,915 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated default.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,915 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated priority.RWQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=2, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,915 INFO [Time-limited test {}] ipc.RWQueueRpcExecutor(113): priority.RWQ.Fifo writeQueues=1 writeHandlers=1 readQueues=1 readHandlers=2 scanQueues=0 scanHandlers=0 2025-01-08T02:20:08,915 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated replication.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=3 2025-01-08T02:20:08,915 INFO [Time-limited test {}] ipc.RpcExecutor(188): Instantiated metaPriority.FPBQ.Fifo with queueClass=class java.util.concurrent.LinkedBlockingQueue; numCallQueues=1, maxQueueLength=30, handlerCount=1 2025-01-08T02:20:08,916 INFO [Time-limited test {}] ipc.RpcServerFactory(64): Creating org.apache.hadoop.hbase.ipc.NettyRpcServer hosting hbase.pb.ClientService, hbase.pb.AdminService, hbase.pb.ClientMetaService, hbase.pb.BootstrapNodeService 2025-01-08T02:20:08,916 INFO [Time-limited test {}] ipc.NettyRpcServer(315): Using org.apache.hbase.thirdparty.io.netty.buffer.PooledByteBufAllocator for buffer allocation 2025-01-08T02:20:08,920 INFO [Time-limited test {}] ipc.NettyRpcServer(197): Bind to /172.17.0.3:39911 2025-01-08T02:20:08,921 INFO [Time-limited test {}] hfile.BlockCacheFactory(123): Allocating BlockCache size=880 MB, blockSize=64 KB 2025-01-08T02:20:08,922 INFO [Time-limited test {}] mob.MobFileCache(128): MobFileCache disabled 2025-01-08T02:20:08,924 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,927 INFO [Time-limited test {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:08,931 INFO [Time-limited test {}] zookeeper.RecoverableZooKeeper(138): Process identifier=regionserver:39911 connecting to ZooKeeper ensemble=127.0.0.1:61593 2025-01-08T02:20:08,936 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:399110x0, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null 2025-01-08T02:20:08,937 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:399110x0, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T02:20:08,938 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:399110x0, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:20:08,939 DEBUG [Time-limited test {}] zookeeper.ZKUtil(113): regionserver:399110x0, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/acl 2025-01-08T02:20:08,940 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKWatcher(635): regionserver:39911-0x1015754588a0003 connected 2025-01-08T02:20:08,941 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=default.FPBQ.Fifo, numCallQueues=1, port=39911 2025-01-08T02:20:08,941 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=priority.RWQ.Fifo.write, numCallQueues=1, port=39911 2025-01-08T02:20:08,948 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=priority.RWQ.Fifo.read, numCallQueues=1, port=39911 2025-01-08T02:20:08,953 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=3 with threadPrefix=replication.FPBQ.Fifo, numCallQueues=1, port=39911 2025-01-08T02:20:08,954 DEBUG [Time-limited test {}] ipc.RpcExecutor(290): Started handlerCount=1 with threadPrefix=metaPriority.FPBQ.Fifo, numCallQueues=1, port=39911 2025-01-08T02:20:08,966 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(2460): Adding backup master ZNode /hbase/backup-masters/c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:08,975 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:08,975 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:08,975 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:08,976 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:08,978 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on existing znode=/hbase/backup-masters/c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:08,987 DEBUG [M:0;c4e5d52b9759:41063 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:M:0;c4e5d52b9759:41063 2025-01-08T02:20:09,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T02:20:09,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T02:20:09,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T02:20:09,019 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/master 2025-01-08T02:20:09,020 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,020 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,020 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,022 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(111): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2025-01-08T02:20:09,023 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(111): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on existing znode=/hbase/master 2025-01-08T02:20:09,023 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.ActiveMasterManager(245): Deleting ZNode for /hbase/backup-masters/c4e5d52b9759,41063,1736302807767 from backup master directory 2025-01-08T02:20:09,028 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:09,028 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/backup-masters/c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:09,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:09,034 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:09,035 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/backup-masters 2025-01-08T02:20:09,036 WARN [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T02:20:09,036 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.ActiveMasterManager(255): Registered as active master=c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:09,040 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating data MemStoreChunkPool with chunk size 2 MB, max count 396, initial count 0 2025-01-08T02:20:09,042 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.ChunkCreator(488): Allocating index MemStoreChunkPool with chunk size 204.80 KB, max count 440, initial count 0 2025-01-08T02:20:09,183 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741826_1002 (size=42) 2025-01-08T02:20:09,184 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741826_1002 (size=42) 2025-01-08T02:20:09,187 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741826_1002 (size=42) 2025-01-08T02:20:09,191 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] util.FSUtils(639): Created cluster ID file at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/hbase.id with ID: 4c56beb2-d2e5-4618-a855-4b17772b2774 2025-01-08T02:20:09,254 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] fs.HFileSystem(339): Added intercepting call to namenode#getBlockLocations so can do block reordering using class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks 2025-01-08T02:20:09,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,301 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:09,346 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741827_1003 (size=196) 2025-01-08T02:20:09,346 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741827_1003 (size=196) 2025-01-08T02:20:09,348 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741827_1003 (size=196) 2025-01-08T02:20:09,384 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] region.MasterRegion(372): Create or load local region for table 'master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:20:09,387 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(132): Injected flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000 2025-01-08T02:20:09,413 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(244): No decryptEncryptedDataEncryptionKey method in DFSClient, should be hadoop version with HDFS-12396 java.lang.NoSuchMethodException: org.apache.hadoop.hdfs.DFSClient.decryptEncryptedDataEncryptionKey(org.apache.hadoop.fs.FileEncryptionInfo) at java.lang.Class.getDeclaredMethod(Class.java:2675) ~[?:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelperWithoutHDFS12396(FanOutOneBlockAsyncDFSOutputSaslHelper.java:183) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.createTransparentCryptoHelper(FanOutOneBlockAsyncDFSOutputSaslHelper.java:242) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper.(FanOutOneBlockAsyncDFSOutputSaslHelper.java:253) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at java.lang.Class.forName0(Native Method) ~[?:?] at java.lang.Class.forName(Class.java:375) ~[?:?] at org.apache.hadoop.hbase.wal.AsyncFSWALProvider.load(AsyncFSWALProvider.java:147) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProviderClass(WALFactory.java:160) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.getProvider(WALFactory.java:200) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:232) ~[classes/:?] at org.apache.hadoop.hbase.wal.WALFactory.(WALFactory.java:207) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.create(MasterRegion.java:402) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegionFactory.create(MasterRegionFactory.java:135) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.finishActiveMasterInitialization(HMaster.java:984) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.startActiveMasterManager(HMaster.java:2485) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.lambda$run$0(HMaster.java:601) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:177) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.master.HMaster.lambda$run$1(HMaster.java:598) ~[classes/:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:20:09,421 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T02:20:09,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741828_1004 (size=1189) 2025-01-08T02:20:09,528 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741828_1004 (size=1189) 2025-01-08T02:20:09,528 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741828_1004 (size=1189) 2025-01-08T02:20:09,558 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(7124): Creating {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='master:store', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'proc', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'rs', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'state', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, under table dir hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/data/master/store 2025-01-08T02:20:09,584 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741829_1005 (size=34) 2025-01-08T02:20:09,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741829_1005 (size=34) 2025-01-08T02:20:09,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741829_1005 (size=34) 2025-01-08T02:20:09,596 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] throttle.StoreHotnessProtector(112): StoreHotnessProtector is disabled. Set hbase.region.store.parallel.put.limit > 0 to enable, which may help mitigate load under heavy write pressure. 2025-01-08T02:20:09,597 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:09,598 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2025-01-08T02:20:09,598 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T02:20:09,599 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T02:20:09,599 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2025-01-08T02:20:09,599 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T02:20:09,599 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1922): Closed master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T02:20:09,599 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T02:20:09,601 WARN [master/c4e5d52b9759:0:becomeActiveMaster {}] region.MasterRegion(249): failed to clean up initializing flag: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/data/master/store/.initializing 2025-01-08T02:20:09,601 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] region.MasterRegion(219): WALDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:09,613 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T02:20:09,637 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=c4e5d52b9759%2C41063%2C1736302807767, suffix=, logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767, archiveDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/oldWALs, maxLogs=10 2025-01-08T02:20:09,672 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644, exclude list is [], retry=0 2025-01-08T02:20:09,690 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41199,DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e,DISK] 2025-01-08T02:20:09,690 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43235,DS-ca0c33db-d569-4a43-bb83-bf5f6c1dccb8,DISK] 2025-01-08T02:20:09,690 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41871,DS-c2bc8ca6-8220-462d-8658-7119831cd0d8,DISK] 2025-01-08T02:20:09,694 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.ProtobufDecoder(117): Hadoop 3.3 and above shades protobuf. 2025-01-08T02:20:09,735 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644 2025-01-08T02:20:09,737 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:45919:45919),(127.0.0.1/127.0.0.1:41865:41865),(127.0.0.1/127.0.0.1:39123:39123)] 2025-01-08T02:20:09,738 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(7285): Opening region: {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''} 2025-01-08T02:20:09,738 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(894): Instantiated master:store,,1.1595e783b53d99cd5eef43b6debb2682.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:09,742 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(7327): checking encryption for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,743 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(7330): checking classloading for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,799 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,831 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName info 2025-01-08T02:20:09,837 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:09,840 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T02:20:09,844 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family proc of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,848 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName proc 2025-01-08T02:20:09,848 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:09,850 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/proc, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:09,850 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rs of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,855 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName rs 2025-01-08T02:20:09,856 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:09,857 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/rs, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:09,857 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family state of region 1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,860 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1595e783b53d99cd5eef43b6debb2682 columnFamilyName state 2025-01-08T02:20:09,861 DEBUG [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:09,863 INFO [StoreOpener-1595e783b53d99cd5eef43b6debb2682-1 {}] regionserver.HStore(327): Store=1595e783b53d99cd5eef43b6debb2682/state, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:09,868 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,871 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,895 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table master:store descriptor;using region.getMemStoreFlushHeapSize/# of families (32.0 M)) instead. 2025-01-08T02:20:09,904 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1085): writing seq id for 1595e783b53d99cd5eef43b6debb2682 2025-01-08T02:20:09,910 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/data/master/store/1595e783b53d99cd5eef43b6debb2682/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:09,911 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1102): Opened 1595e783b53d99cd5eef43b6debb2682; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=75105427, jitterRate=0.119158074259758}}}, FlushLargeStoresPolicy{flushSizeLowerBound=33554432} 2025-01-08T02:20:09,915 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] regionserver.HRegion(1001): Region open journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T02:20:09,916 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] region.MasterRegionFlusherAndCompactor(122): Constructor flushSize=134217728, flushPerChanges=1000000, flushIntervalMs=900000, compactMin=4 2025-01-08T02:20:09,947 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3bcb2cc1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:09,987 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(893): No meta location available on zookeeper, skip migrating... 2025-01-08T02:20:10,000 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] region.RegionProcedureStore(104): Starting the Region Procedure Store, number threads=5 2025-01-08T02:20:10,001 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(633): Starting 5 core workers (bigger of cpus/4 or 16) with max (burst) worker count=50 2025-01-08T02:20:10,004 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] region.RegionProcedureStore(255): Starting Region Procedure Store lease recovery... 2025-01-08T02:20:10,005 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(653): Recovered RegionProcedureStore lease in 1 msec 2025-01-08T02:20:10,011 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(667): Loaded RegionProcedureStore in 5 msec 2025-01-08T02:20:10,011 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.RemoteProcedureDispatcher(96): Instantiated, coreThreads=3 (allowCoreThreadTimeOut=true), queueMaxSize=32, operationDelay=150 2025-01-08T02:20:10,045 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.RegionServerTracker(127): Upgrading RegionServerTracker to active master mode; 0 have existingServerCrashProcedures, 0 possibly 'live' servers, and 0 'splitting'. 2025-01-08T02:20:10,058 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Unable to get data of znode /hbase/balancer because node does not exist (not necessarily an error) 2025-01-08T02:20:10,061 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/balancer already deleted, retry=false 2025-01-08T02:20:10,064 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] normalizer.SimpleRegionNormalizer(163): Updated configuration for key 'hbase.normalizer.merge.min_region_size.mb' from 0 to 1 2025-01-08T02:20:10,065 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Unable to get data of znode /hbase/normalizer because node does not exist (not necessarily an error) 2025-01-08T02:20:10,068 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/normalizer already deleted, retry=false 2025-01-08T02:20:10,070 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] normalizer.RegionNormalizerWorker(137): Normalizer rate limit set to unlimited 2025-01-08T02:20:10,074 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Unable to get data of znode /hbase/switch/split because node does not exist (not necessarily an error) 2025-01-08T02:20:10,076 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/split already deleted, retry=false 2025-01-08T02:20:10,077 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Unable to get data of znode /hbase/switch/merge because node does not exist (not necessarily an error) 2025-01-08T02:20:10,079 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/switch/merge already deleted, retry=false 2025-01-08T02:20:10,092 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(444): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Unable to get data of znode /hbase/snapshot-cleanup because node does not exist (not necessarily an error) 2025-01-08T02:20:10,094 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/snapshot-cleanup already deleted, retry=false 2025-01-08T02:20:10,098 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T02:20:10,098 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T02:20:10,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T02:20:10,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/running 2025-01-08T02:20:10,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,099 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,100 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(837): Active/primary master=c4e5d52b9759,41063,1736302807767, sessionid=0x1015754588a0000, setting cluster-up flag (Was=false) 2025-01-08T02:20:10,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,123 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,152 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/flush-table-proc/acquired, /hbase/flush-table-proc/reached, /hbase/flush-table-proc/abort 2025-01-08T02:20:10,154 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:10,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,170 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:10,180 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure.ZKProcedureUtil(251): Clearing all znodes /hbase/online-snapshot/acquired, /hbase/online-snapshot/reached, /hbase/online-snapshot/abort 2025-01-08T02:20:10,184 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure.ZKProcedureCoordinator(245): Starting controller for procedure member=c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:10,301 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:0;c4e5d52b9759:37031 2025-01-08T02:20:10,304 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1008): ClusterId : 4c56beb2-d2e5-4618-a855-4b17772b2774 2025-01-08T02:20:10,309 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T02:20:10,319 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:1;c4e5d52b9759:36743 2025-01-08T02:20:10,321 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T02:20:10,324 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T02:20:10,325 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1008): ClusterId : 4c56beb2-d2e5-4618-a855-4b17772b2774 2025-01-08T02:20:10,325 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T02:20:10,329 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T02:20:10,331 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.ShutdownHook(81): Installed shutdown hook thread: Shutdownhook:RS:2;c4e5d52b9759:39911 2025-01-08T02:20:10,332 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T02:20:10,332 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T02:20:10,332 DEBUG [RS:0;c4e5d52b9759:37031 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4193618f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:10,333 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1008): ClusterId : 4c56beb2-d2e5-4618-a855-4b17772b2774 2025-01-08T02:20:10,333 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(43): Procedure flush-table-proc initializing 2025-01-08T02:20:10,337 DEBUG [RS:0;c4e5d52b9759:37031 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@742f0bc6, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=c4e5d52b9759/172.17.0.3:0 2025-01-08T02:20:10,338 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T02:20:10,341 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(45): Procedure flush-table-proc initialized 2025-01-08T02:20:10,341 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(43): Procedure online-snapshot initializing 2025-01-08T02:20:10,344 DEBUG [RS:1;c4e5d52b9759:36743 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@16c59174, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:10,344 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(45): Procedure online-snapshot initialized 2025-01-08T02:20:10,345 DEBUG [RS:2;c4e5d52b9759:39911 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@51996f93, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:10,345 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T02:20:10,345 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T02:20:10,346 DEBUG [RS:1;c4e5d52b9759:36743 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@10889b83, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=c4e5d52b9759/172.17.0.3:0 2025-01-08T02:20:10,347 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T02:20:10,347 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T02:20:10,349 DEBUG [RS:2;c4e5d52b9759:39911 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7672afc0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=c4e5d52b9759/172.17.0.3:0 2025-01-08T02:20:10,350 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.RegionServerCoprocessorHost(67): System coprocessor loading is enabled 2025-01-08T02:20:10,350 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.RegionServerCoprocessorHost(68): Table coprocessor loading is enabled 2025-01-08T02:20:10,357 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T02:20:10,357 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T02:20:10,357 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(900): Registered regionserver coprocessor executorService: executorService=AccessControlService 2025-01-08T02:20:10,357 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(3405): Registered master coprocessor service: service=AccessControlService 2025-01-08T02:20:10,358 INFO [RS:1;c4e5d52b9759:36743 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:10,358 INFO [RS:0;c4e5d52b9759:37031 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:10,358 INFO [RS:2;c4e5d52b9759:39911 {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:10,358 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T02:20:10,358 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T02:20:10,358 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1090): About to register with Master. 2025-01-08T02:20:10,358 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:10,359 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver loaded, priority=536870912. 2025-01-08T02:20:10,362 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(3073): reportForDuty to master=c4e5d52b9759,41063,1736302807767 with isa=c4e5d52b9759/172.17.0.3:36743, startcode=1736302808837 2025-01-08T02:20:10,362 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(3073): reportForDuty to master=c4e5d52b9759,41063,1736302807767 with isa=c4e5d52b9759/172.17.0.3:39911, startcode=1736302808913 2025-01-08T02:20:10,362 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(3073): reportForDuty to master=c4e5d52b9759,41063,1736302807767 with isa=c4e5d52b9759/172.17.0.3:37031, startcode=1736302808764 2025-01-08T02:20:10,387 DEBUG [RS:1;c4e5d52b9759:36743 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T02:20:10,387 DEBUG [RS:2;c4e5d52b9759:39911 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T02:20:10,387 DEBUG [RS:0;c4e5d52b9759:37031 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T02:20:10,451 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT; InitMetaProcedure table=hbase:meta 2025-01-08T02:20:10,459 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:47149, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T02:20:10,459 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54093, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T02:20:10,459 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:52435, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T02:20:10,460 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] balancer.BaseLoadBalancer(575): slop=0.2 2025-01-08T02:20:10,463 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] balancer.StochasticLoadBalancer(294): Loaded config; maxSteps=1000000, runMaxSteps=false, stepsPerRegion=800, maxRunningTime=30000, isByTable=false, CostFunctions=[RegionCountSkewCostFunction, PrimaryRegionCountSkewCostFunction, MoveCostFunction, ServerLocalityCostFunction, RackLocalityCostFunction, TableSkewCostFunction, RegionReplicaHostCostFunction, RegionReplicaRackCostFunction, ReadRequestCostFunction, WriteRequestCostFunction, MemStoreSizeCostFunction, StoreFileCostFunction] , sum of multiplier of cost functions = 0.0 etc. 2025-01-08T02:20:10,467 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T02:20:10,472 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] balancer.RegionLocationFinder(146): Skipping locality-based refresh due to oldStatus=null, newStatus=Master: c4e5d52b9759,41063,1736302807767 Number of backup masters: 0 Number of live region servers: 0 Number of dead region servers: 0 Number of unknown region servers: 0 Average load: 0.0 Number of requests: 0 Number of regions: 0 Number of regions in transition: 0 2025-01-08T02:20:10,475 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_OPEN_REGION-master/c4e5d52b9759:0, corePoolSize=5, maxPoolSize=5 2025-01-08T02:20:10,476 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_CLOSE_REGION-master/c4e5d52b9759:0, corePoolSize=5, maxPoolSize=5 2025-01-08T02:20:10,476 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SERVER_OPERATIONS-master/c4e5d52b9759:0, corePoolSize=5, maxPoolSize=5 2025-01-08T02:20:10,476 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_META_SERVER_OPERATIONS-master/c4e5d52b9759:0, corePoolSize=5, maxPoolSize=5 2025-01-08T02:20:10,476 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=M_LOG_REPLAY_OPS-master/c4e5d52b9759:0, corePoolSize=10, maxPoolSize=10 2025-01-08T02:20:10,476 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_SNAPSHOT_OPERATIONS-master/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,477 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_MERGE_OPERATIONS-master/c4e5d52b9759:0, corePoolSize=2, maxPoolSize=2 2025-01-08T02:20:10,477 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] executor.ExecutorService(95): Starting executor service name=MASTER_TABLE_OPERATIONS-master/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,477 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T02:20:10,478 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] ipc.MetricsHBaseServer(152): Unknown exception type org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:3295) ~[classes/:?] at org.apache.hadoop.hbase.master.MasterRpcServices.regionServerStartup(MasterRpcServices.java:593) ~[classes/:?] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$2.callBlockingMethod(RegionServerStatusProtos.java:16714) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:443) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:124) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:105) ~[classes/:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) ~[classes/:2.7.0-SNAPSHOT] 2025-01-08T02:20:10,486 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.procedure2.CompletedProcedureCleaner; timeout=30000, timestamp=1736302840486 2025-01-08T02:20:10,487 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_WRITE_FS_LAYOUT, locked=true; InitMetaProcedure table=hbase:meta 2025-01-08T02:20:10,487 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(75): BOOTSTRAP: creating hbase:meta region 2025-01-08T02:20:10,488 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.DirScanPool(74): log_cleaner Cleaner pool size is 1 2025-01-08T02:20:10,489 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner 2025-01-08T02:20:10,492 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:10,492 INFO [PEWorker-1 {}] util.FSTableDescriptors(133): Creating new hbase:meta table descriptor 'hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T02:20:10,493 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.replication.master.ReplicationLogCleaner 2025-01-08T02:20:10,493 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreWALCleaner 2025-01-08T02:20:10,493 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner 2025-01-08T02:20:10,493 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.LogCleaner(148): Creating 1 old WALs cleaner threads 2025-01-08T02:20:10,499 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=LogsCleaner, period=600000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,502 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.DirScanPool(74): hfile_cleaner Cleaner pool size is 2 2025-01-08T02:20:10,503 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T02:20:10,503 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T02:20:10,503 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(3097): Master is not running yet 2025-01-08T02:20:10,503 WARN [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T02:20:10,503 WARN [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T02:20:10,503 WARN [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1099): reportForDuty failed; sleeping 100 ms and then retrying. 2025-01-08T02:20:10,504 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveMasterLocalStoreHFileCleaner 2025-01-08T02:20:10,505 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.HFileLinkCleaner 2025-01-08T02:20:10,507 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.snapshot.SnapshotHFileCleaner 2025-01-08T02:20:10,507 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.CleanerChore(192): Initialize cleaner=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner 2025-01-08T02:20:10,513 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.HFileCleaner(260): Starting for large file=Thread[master/c4e5d52b9759:0:becomeActiveMaster-HFileCleaner.large.0-1736302810509,5,FailOnTimeoutGroup] 2025-01-08T02:20:10,513 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] cleaner.HFileCleaner(275): Starting for small files=Thread[master/c4e5d52b9759:0:becomeActiveMaster-HFileCleaner.small.0-1736302810513,5,FailOnTimeoutGroup] 2025-01-08T02:20:10,513 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HFileCleaner, period=600000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,514 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(1691): Reopening regions with very high storeFileRefCount is disabled. Provide threshold value > 0 for hbase.regions.recovery.store.file.ref.count to enable it. 2025-01-08T02:20:10,515 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=ReplicationBarrierCleaner, period=43200000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,515 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=SnapshotCleaner, period=1800000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,516 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741831_1007 (size=1039) 2025-01-08T02:20:10,517 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741831_1007 (size=1039) 2025-01-08T02:20:10,518 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741831_1007 (size=1039) 2025-01-08T02:20:10,519 INFO [PEWorker-1 {}] util.FSTableDescriptors(140): Updated hbase:meta table descriptor to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/.tabledesc/.tableinfo.0000000001.1039 2025-01-08T02:20:10,520 INFO [PEWorker-1 {}] regionserver.HRegion(7106): creating {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:meta', {TABLE_ATTRIBUTES => {IS_META => 'true', coprocessor$1 => '|org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint|536870911|', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, {NAME => 'rep_barrier', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '2147483647', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, {NAME => 'table', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '3', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'ROW_INDEX_V1', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROWCOL', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:10,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741832_1008 (size=32) 2025-01-08T02:20:10,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741832_1008 (size=32) 2025-01-08T02:20:10,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741832_1008 (size=32) 2025-01-08T02:20:10,541 DEBUG [PEWorker-1 {}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:10,544 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2025-01-08T02:20:10,547 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2025-01-08T02:20:10,547 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:10,548 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T02:20:10,548 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2025-01-08T02:20:10,550 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2025-01-08T02:20:10,550 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:10,551 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T02:20:10,551 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2025-01-08T02:20:10,554 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2025-01-08T02:20:10,554 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:10,555 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T02:20:10,556 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740 2025-01-08T02:20:10,557 DEBUG [PEWorker-1 {}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740 2025-01-08T02:20:10,560 DEBUG [PEWorker-1 {}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2025-01-08T02:20:10,562 DEBUG [PEWorker-1 {}] regionserver.HRegion(1085): writing seq id for 1588230740 2025-01-08T02:20:10,567 DEBUG [PEWorker-1 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:10,568 INFO [PEWorker-1 {}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61546920, jitterRate=-0.08287942409515381}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2025-01-08T02:20:10,571 DEBUG [PEWorker-1 {}] regionserver.HRegion(1001): Region open journal for 1588230740: 2025-01-08T02:20:10,572 DEBUG [PEWorker-1 {}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2025-01-08T02:20:10,572 INFO [PEWorker-1 {}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2025-01-08T02:20:10,572 DEBUG [PEWorker-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2025-01-08T02:20:10,572 DEBUG [PEWorker-1 {}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2025-01-08T02:20:10,572 DEBUG [PEWorker-1 {}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2025-01-08T02:20:10,574 INFO [PEWorker-1 {}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2025-01-08T02:20:10,574 DEBUG [PEWorker-1 {}] regionserver.HRegion(1635): Region close journal for 1588230740: 2025-01-08T02:20:10,577 DEBUG [PEWorker-1 {}] procedure.InitMetaProcedure(96): Execute pid=1, state=RUNNABLE:INIT_META_ASSIGN_META, locked=true; InitMetaProcedure table=hbase:meta 2025-01-08T02:20:10,577 INFO [PEWorker-1 {}] procedure.InitMetaProcedure(107): Going to assign meta 2025-01-08T02:20:10,584 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN}] 2025-01-08T02:20:10,594 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN 2025-01-08T02:20:10,599 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=2, ppid=1, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN; state=OFFLINE, location=null; forceNewPlan=false, retain=false 2025-01-08T02:20:10,605 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(3073): reportForDuty to master=c4e5d52b9759,41063,1736302807767 with isa=c4e5d52b9759/172.17.0.3:37031, startcode=1736302808764 2025-01-08T02:20:10,605 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(3073): reportForDuty to master=c4e5d52b9759,41063,1736302807767 with isa=c4e5d52b9759/172.17.0.3:36743, startcode=1736302808837 2025-01-08T02:20:10,605 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(3073): reportForDuty to master=c4e5d52b9759,41063,1736302807767 with isa=c4e5d52b9759/172.17.0.3:39911, startcode=1736302808913 2025-01-08T02:20:10,606 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] master.ServerManager(332): Checking decommissioned status of RegionServer c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:10,609 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] master.ServerManager(486): Registering regionserver=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:10,618 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] master.ServerManager(332): Checking decommissioned status of RegionServer c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:10,618 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] master.ServerManager(486): Registering regionserver=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:10,618 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:10,618 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:43527 2025-01-08T02:20:10,618 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T02:20:10,620 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] master.ServerManager(332): Checking decommissioned status of RegionServer c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:10,620 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] master.ServerManager(486): Registering regionserver=c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:10,620 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:10,621 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:43527 2025-01-08T02:20:10,621 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T02:20:10,623 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T02:20:10,623 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1725): Config from master: hbase.rootdir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:10,623 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1725): Config from master: fs.defaultFS=hdfs://localhost:43527 2025-01-08T02:20:10,623 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1725): Config from master: hbase.master.info.port=-1 2025-01-08T02:20:10,627 DEBUG [RS:0;c4e5d52b9759:37031 {}] zookeeper.ZKUtil(111): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:10,627 WARN [RS:0;c4e5d52b9759:37031 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T02:20:10,627 INFO [RS:0;c4e5d52b9759:37031 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T02:20:10,628 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:10,629 DEBUG [RS:1;c4e5d52b9759:36743 {}] zookeeper.ZKUtil(111): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:10,629 WARN [RS:1;c4e5d52b9759:36743 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T02:20:10,629 DEBUG [RS:2;c4e5d52b9759:39911 {}] zookeeper.ZKUtil(111): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on existing znode=/hbase/rs/c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:10,629 INFO [RS:1;c4e5d52b9759:36743 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T02:20:10,629 WARN [RS:2;c4e5d52b9759:39911 {}] hbase.ZNodeClearer(69): Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared on crash by start scripts (Longer MTTR!) 2025-01-08T02:20:10,629 INFO [RS:2;c4e5d52b9759:39911 {}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T02:20:10,629 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:10,630 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(2100): logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:10,630 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [c4e5d52b9759,37031,1736302808764] 2025-01-08T02:20:10,631 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [c4e5d52b9759,36743,1736302808837] 2025-01-08T02:20:10,631 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(179): RegionServer ephemeral node created, adding [c4e5d52b9759,39911,1736302808913] 2025-01-08T02:20:10,646 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T02:20:10,646 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T02:20:10,646 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.Replication(140): Replication stats-in-log period=300 seconds 2025-01-08T02:20:10,658 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T02:20:10,658 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T02:20:10,658 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.MetricsRegionServerWrapperImpl(120): Computing regionserver metrics every 5000 milliseconds 2025-01-08T02:20:10,671 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T02:20:10,671 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T02:20:10,671 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.MemStoreFlusher(130): globalMemStoreLimit=880 M, globalMemStoreLimitLowMark=836 M, Offheap=false 2025-01-08T02:20:10,674 INFO [RS:0;c4e5d52b9759:37031 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T02:20:10,674 INFO [RS:1;c4e5d52b9759:36743 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T02:20:10,674 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,674 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,676 INFO [RS:2;c4e5d52b9759:39911 {}] throttle.PressureAwareCompactionThroughputController(131): Compaction throughput configurations, higher bound: 100.00 MB/second, lower bound 50.00 MB/second, off peak: unlimited, tuning period: 60000 ms 2025-01-08T02:20:10,676 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,676 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T02:20:10,676 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T02:20:10,676 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer$CompactionChecker(1988): CompactionChecker runs every PT1S 2025-01-08T02:20:10,684 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,684 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,684 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactedHFilesCleaner, period=120000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,684 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,684 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,684 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,684 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,684 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,684 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_META-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_REGION-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_CLOSE_META-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/c4e5d52b9759:0, corePoolSize=2, maxPoolSize=2 2025-01-08T02:20:10,685 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/c4e5d52b9759:0, corePoolSize=2, maxPoolSize=2 2025-01-08T02:20:10,685 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_LOG_REPLAY_OPS-regionserver/c4e5d52b9759:0, corePoolSize=2, maxPoolSize=2 2025-01-08T02:20:10,685 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_COMPACTED_FILES_DISCHARGER-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,685 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_REGION_REPLICA_FLUSH_OPS-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_REFRESH_PEER-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_SWITCH_RPC_THROTTLE-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_CLAIM_REPLICATION_QUEUE-regionserver/c4e5d52b9759:0, corePoolSize=1, maxPoolSize=1 2025-01-08T02:20:10,686 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0, corePoolSize=3, maxPoolSize=3 2025-01-08T02:20:10,686 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0, corePoolSize=3, maxPoolSize=3 2025-01-08T02:20:10,686 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0, corePoolSize=3, maxPoolSize=3 2025-01-08T02:20:10,686 DEBUG [RS:0;c4e5d52b9759:37031 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/c4e5d52b9759:0, corePoolSize=3, maxPoolSize=3 2025-01-08T02:20:10,686 DEBUG [RS:1;c4e5d52b9759:36743 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/c4e5d52b9759:0, corePoolSize=3, maxPoolSize=3 2025-01-08T02:20:10,686 DEBUG [RS:2;c4e5d52b9759:39911 {}] executor.ExecutorService(95): Starting executor service name=RS_FLUSH_OPERATIONS-regionserver/c4e5d52b9759:0, corePoolSize=3, maxPoolSize=3 2025-01-08T02:20:10,690 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=CompactionChecker, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=MemstoreFlusherChore, period=1000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,37031,1736302808764-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=nonceCleaner, period=360000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,36743,1736302808837-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T02:20:10,690 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,691 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,39911,1736302808913-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T02:20:10,712 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T02:20:10,713 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T02:20:10,714 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,36743,1736302808837-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,714 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,37031,1736302808764-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,718 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HeapMemoryManager(209): Starting, tuneOn=false 2025-01-08T02:20:10,718 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,39911,1736302808913-HeapMemoryTunerChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:10,736 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.Replication(204): c4e5d52b9759,36743,1736302808837 started 2025-01-08T02:20:10,736 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.Replication(204): c4e5d52b9759,37031,1736302808764 started 2025-01-08T02:20:10,736 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1767): Serving as c4e5d52b9759,36743,1736302808837, RpcServer on c4e5d52b9759/172.17.0.3:36743, sessionid=0x1015754588a0002 2025-01-08T02:20:10,736 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1767): Serving as c4e5d52b9759,37031,1736302808764, RpcServer on c4e5d52b9759/172.17.0.3:37031, sessionid=0x1015754588a0001 2025-01-08T02:20:10,736 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T02:20:10,736 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T02:20:10,736 DEBUG [RS:1;c4e5d52b9759:36743 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:10,736 DEBUG [RS:0;c4e5d52b9759:37031 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:10,737 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'c4e5d52b9759,36743,1736302808837' 2025-01-08T02:20:10,737 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'c4e5d52b9759,37031,1736302808764' 2025-01-08T02:20:10,737 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T02:20:10,737 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T02:20:10,737 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T02:20:10,738 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T02:20:10,738 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T02:20:10,738 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T02:20:10,738 DEBUG [RS:1;c4e5d52b9759:36743 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:10,739 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'c4e5d52b9759,36743,1736302808837' 2025-01-08T02:20:10,739 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T02:20:10,740 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T02:20:10,740 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T02:20:10,740 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T02:20:10,740 DEBUG [RS:0;c4e5d52b9759:37031 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:10,740 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'c4e5d52b9759,37031,1736302808764' 2025-01-08T02:20:10,740 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T02:20:10,740 DEBUG [RS:1;c4e5d52b9759:36743 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T02:20:10,741 INFO [RS:1;c4e5d52b9759:36743 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T02:20:10,741 INFO [RS:1;c4e5d52b9759:36743 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T02:20:10,741 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T02:20:10,741 DEBUG [RS:0;c4e5d52b9759:37031 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T02:20:10,741 INFO [RS:0;c4e5d52b9759:37031 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T02:20:10,742 INFO [RS:0;c4e5d52b9759:37031 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T02:20:10,742 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.Replication(204): c4e5d52b9759,39911,1736302808913 started 2025-01-08T02:20:10,743 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1767): Serving as c4e5d52b9759,39911,1736302808913, RpcServer on c4e5d52b9759/172.17.0.3:39911, sessionid=0x1015754588a0003 2025-01-08T02:20:10,743 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(51): Procedure flush-table-proc starting 2025-01-08T02:20:10,743 DEBUG [RS:2;c4e5d52b9759:39911 {}] flush.RegionServerFlushTableProcedureManager(108): Start region server flush procedure manager c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:10,743 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'c4e5d52b9759,39911,1736302808913' 2025-01-08T02:20:10,743 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/flush-table-proc/abort' 2025-01-08T02:20:10,743 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/flush-table-proc/acquired' 2025-01-08T02:20:10,745 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(53): Procedure flush-table-proc started 2025-01-08T02:20:10,745 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(51): Procedure online-snapshot starting 2025-01-08T02:20:10,745 DEBUG [RS:2;c4e5d52b9759:39911 {}] snapshot.RegionServerSnapshotManager(126): Start Snapshot Manager c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:10,745 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.ZKProcedureMemberRpcs(357): Starting procedure member 'c4e5d52b9759,39911,1736302808913' 2025-01-08T02:20:10,745 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.ZKProcedureMemberRpcs(134): Checking for aborted procedures on node: '/hbase/online-snapshot/abort' 2025-01-08T02:20:10,745 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.ZKProcedureMemberRpcs(154): Looking for new procedures under znode:'/hbase/online-snapshot/acquired' 2025-01-08T02:20:10,747 DEBUG [RS:2;c4e5d52b9759:39911 {}] procedure.RegionServerProcedureManagerHost(53): Procedure online-snapshot started 2025-01-08T02:20:10,747 INFO [RS:2;c4e5d52b9759:39911 {}] quotas.RegionServerRpcQuotaManager(64): Quota support disabled 2025-01-08T02:20:10,747 INFO [RS:2;c4e5d52b9759:39911 {}] quotas.RegionServerSpaceQuotaManager(80): Quota support disabled, not starting space quota manager. 2025-01-08T02:20:10,750 WARN [c4e5d52b9759:41063 {}] assignment.AssignmentManager(2423): No servers available; cannot place 1 unassigned regions. 2025-01-08T02:20:10,846 INFO [RS:1;c4e5d52b9759:36743 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T02:20:10,846 INFO [RS:0;c4e5d52b9759:37031 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T02:20:10,847 INFO [RS:2;c4e5d52b9759:39911 {}] monitor.StreamSlowMonitor(122): New stream slow monitor defaultMonitorName 2025-01-08T02:20:10,850 INFO [RS:1;c4e5d52b9759:36743 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=c4e5d52b9759%2C36743%2C1736302808837, suffix=, logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,36743,1736302808837, archiveDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs, maxLogs=32 2025-01-08T02:20:10,850 INFO [RS:0;c4e5d52b9759:37031 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=c4e5d52b9759%2C37031%2C1736302808764, suffix=, logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,37031,1736302808764, archiveDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs, maxLogs=32 2025-01-08T02:20:10,850 INFO [RS:2;c4e5d52b9759:39911 {}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=c4e5d52b9759%2C39911%2C1736302808913, suffix=, logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,39911,1736302808913, archiveDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs, maxLogs=32 2025-01-08T02:20:10,869 DEBUG [RS:1;c4e5d52b9759:36743 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,36743,1736302808837/c4e5d52b9759%2C36743%2C1736302808837.1736302810854, exclude list is [], retry=0 2025-01-08T02:20:10,875 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43235,DS-ca0c33db-d569-4a43-bb83-bf5f6c1dccb8,DISK] 2025-01-08T02:20:10,875 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41199,DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e,DISK] 2025-01-08T02:20:10,875 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41871,DS-c2bc8ca6-8220-462d-8658-7119831cd0d8,DISK] 2025-01-08T02:20:10,876 DEBUG [RS:0;c4e5d52b9759:37031 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,37031,1736302808764/c4e5d52b9759%2C37031%2C1736302808764.1736302810855, exclude list is [], retry=0 2025-01-08T02:20:10,876 DEBUG [RS:2;c4e5d52b9759:39911 {}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,39911,1736302808913/c4e5d52b9759%2C39911%2C1736302808913.1736302810855, exclude list is [], retry=0 2025-01-08T02:20:10,881 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41199,DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e,DISK] 2025-01-08T02:20:10,881 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43235,DS-ca0c33db-d569-4a43-bb83-bf5f6c1dccb8,DISK] 2025-01-08T02:20:10,883 INFO [RS:1;c4e5d52b9759:36743 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,36743,1736302808837/c4e5d52b9759%2C36743%2C1736302808837.1736302810854 2025-01-08T02:20:10,883 DEBUG [RS:1;c4e5d52b9759:36743 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:41865:41865),(127.0.0.1/127.0.0.1:39123:39123),(127.0.0.1/127.0.0.1:45919:45919)] 2025-01-08T02:20:10,905 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41871,DS-c2bc8ca6-8220-462d-8658-7119831cd0d8,DISK] 2025-01-08T02:20:10,906 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41199,DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e,DISK] 2025-01-08T02:20:10,907 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41871,DS-c2bc8ca6-8220-462d-8658-7119831cd0d8,DISK] 2025-01-08T02:20:10,907 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43235,DS-ca0c33db-d569-4a43-bb83-bf5f6c1dccb8,DISK] 2025-01-08T02:20:10,913 INFO [RS:0;c4e5d52b9759:37031 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,37031,1736302808764/c4e5d52b9759%2C37031%2C1736302808764.1736302810855 2025-01-08T02:20:10,914 DEBUG [RS:0;c4e5d52b9759:37031 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:41865:41865),(127.0.0.1/127.0.0.1:45919:45919),(127.0.0.1/127.0.0.1:39123:39123)] 2025-01-08T02:20:10,916 INFO [RS:2;c4e5d52b9759:39911 {}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,39911,1736302808913/c4e5d52b9759%2C39911%2C1736302808913.1736302810855 2025-01-08T02:20:10,920 DEBUG [RS:2;c4e5d52b9759:39911 {}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:41865:41865),(127.0.0.1/127.0.0.1:45919:45919),(127.0.0.1/127.0.0.1:39123:39123)] 2025-01-08T02:20:11,002 DEBUG [c4e5d52b9759:41063 {}] assignment.AssignmentManager(2444): Processing assignQueue; systemServersCount=3, allServersCount=3 2025-01-08T02:20:11,005 DEBUG [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:20:11,012 DEBUG [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:20:11,012 DEBUG [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:20:11,012 DEBUG [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:20:11,012 INFO [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:20:11,012 INFO [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:20:11,012 INFO [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:20:11,012 DEBUG [c4e5d52b9759:41063 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:20:11,018 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:11,023 INFO [PEWorker-3 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as c4e5d52b9759,36743,1736302808837, state=OPENING 2025-01-08T02:20:11,028 DEBUG [PEWorker-3 {}] zookeeper.MetaTableLocator(183): hbase:meta region location doesn't exist, create it 2025-01-08T02:20:11,030 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:11,030 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:11,030 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:11,030 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:11,030 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,030 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,030 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,031 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,032 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=3, ppid=2, state=RUNNABLE; OpenRegionProcedure 1588230740, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:20:11,208 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:11,210 DEBUG [RSProcedureDispatcher-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T02:20:11,213 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33006, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T02:20:11,235 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(135): Open hbase:meta,,1.1588230740 2025-01-08T02:20:11,236 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.WALFactory(183): Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.AsyncFSWALProvider 2025-01-08T02:20:11,236 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] monitor.StreamSlowMonitor(122): New stream slow monitor .meta 2025-01-08T02:20:11,252 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(500): WAL configuration: blocksize=256 MB, rollsize=128 MB, prefix=c4e5d52b9759%2C36743%2C1736302808837.meta, suffix=.meta, logDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,36743,1736302808837, archiveDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs, maxLogs=32 2025-01-08T02:20:11,276 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] asyncfs.FanOutOneBlockAsyncDFSOutputHelper(617): When create output stream for /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,36743,1736302808837/c4e5d52b9759%2C36743%2C1736302808837.meta.1736302811254.meta, exclude list is [], retry=0 2025-01-08T02:20:11,282 DEBUG [RS-EventLoopGroup-5-1 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41199,DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e,DISK] 2025-01-08T02:20:11,282 DEBUG [RS-EventLoopGroup-5-2 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:43235,DS-ca0c33db-d569-4a43-bb83-bf5f6c1dccb8,DISK] 2025-01-08T02:20:11,282 DEBUG [RS-EventLoopGroup-5-3 {}] asyncfs.FanOutOneBlockAsyncDFSOutputSaslHelper(816): SASL client skipping handshake in unsecured configuration for addr = 127.0.0.1/127.0.0.1, datanodeId = DatanodeInfoWithStorage[127.0.0.1:41871,DS-c2bc8ca6-8220-462d-8658-7119831cd0d8,DISK] 2025-01-08T02:20:11,291 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(841): New WAL /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/WALs/c4e5d52b9759,36743,1736302808837/c4e5d52b9759%2C36743%2C1736302808837.meta.1736302811254.meta 2025-01-08T02:20:11,291 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] wal.AbstractFSWAL(925): Create new AsyncFSWAL writer with pipeline: [(127.0.0.1/127.0.0.1:41865:41865),(127.0.0.1/127.0.0.1:45919:45919),(127.0.0.1/127.0.0.1:39123:39123)] 2025-01-08T02:20:11,292 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7285): Opening region: {ENCODED => 1588230740, NAME => 'hbase:meta,,1', STARTKEY => '', ENDKEY => ''} 2025-01-08T02:20:11,293 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=AccessControlService 2025-01-08T02:20:11,296 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:11,297 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] coprocessor.CoprocessorHost(215): Loading coprocessor class org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint with path null and priority 536870911 2025-01-08T02:20:11,298 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:meta,,1 service=MultiRowMutationService 2025-01-08T02:20:11,300 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.RegionCoprocessorHost(436): Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully. 2025-01-08T02:20:11,313 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table meta 1588230740 2025-01-08T02:20:11,314 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(894): Instantiated hbase:meta,,1.1588230740; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:11,314 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7327): checking encryption for 1588230740 2025-01-08T02:20:11,324 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(7330): checking classloading for 1588230740 2025-01-08T02:20:11,327 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 1588230740 2025-01-08T02:20:11,329 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName info 2025-01-08T02:20:11,329 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:11,330 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T02:20:11,330 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family rep_barrier of region 1588230740 2025-01-08T02:20:11,332 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName rep_barrier 2025-01-08T02:20:11,332 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:11,333 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/rep_barrier, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T02:20:11,333 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family table of region 1588230740 2025-01-08T02:20:11,336 INFO [StoreOpener-1588230740-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1588230740 columnFamilyName table 2025-01-08T02:20:11,336 DEBUG [StoreOpener-1588230740-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:11,337 INFO [StoreOpener-1588230740-1 {}] regionserver.HStore(327): Store=1588230740/table, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=ROW_INDEX_V1, compression=NONE 2025-01-08T02:20:11,338 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740 2025-01-08T02:20:11,341 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740 2025-01-08T02:20:11,344 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.FlushLargeStoresPolicy(65): No hbase.hregion.percolumnfamilyflush.size.lower.bound set in table hbase:meta descriptor;using region.getMemStoreFlushHeapSize/# of families (42.7 M)) instead. 2025-01-08T02:20:11,347 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1085): writing seq id for 1588230740 2025-01-08T02:20:11,349 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1102): Opened 1588230740; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63564688, jitterRate=-0.05281233787536621}}}, FlushLargeStoresPolicy{flushSizeLowerBound=44739242} 2025-01-08T02:20:11,353 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegion(1001): Region open journal for 1588230740: 2025-01-08T02:20:11,362 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:meta,,1.1588230740, pid=3, masterSystemTime=1736302811200 2025-01-08T02:20:11,375 DEBUG [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:meta,,1.1588230740 2025-01-08T02:20:11,376 INFO [RS_OPEN_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_META, pid=3}] handler.AssignRegionHandler(164): Opened hbase:meta,,1.1588230740 2025-01-08T02:20:11,377 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=2 updating hbase:meta row=1588230740, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:11,380 INFO [PEWorker-5 {}] zookeeper.MetaTableLocator(171): Setting hbase:meta replicaId=0 location in ZooKeeper as c4e5d52b9759,36743,1736302808837, state=OPEN 2025-01-08T02:20:11,382 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T02:20:11,382 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T02:20:11,383 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,383 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T02:20:11,383 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,383 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/meta-region-server 2025-01-08T02:20:11,383 DEBUG [zk-event-processor-pool-0 {}] hbase.MetaRegionLocationCache(167): Updating meta znode for path /hbase/meta-region-server: CHANGED 2025-01-08T02:20:11,389 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=3, resume processing ppid=2 2025-01-08T02:20:11,389 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=3, ppid=2, state=SUCCESS; OpenRegionProcedure 1588230740, server=c4e5d52b9759,36743,1736302808837 in 351 msec 2025-01-08T02:20:11,395 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=2, resume processing ppid=1 2025-01-08T02:20:11,395 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=2, ppid=1, state=SUCCESS; TransitRegionStateProcedure table=hbase:meta, region=1588230740, ASSIGN in 806 msec 2025-01-08T02:20:11,402 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=1, state=SUCCESS; InitMetaProcedure table=hbase:meta in 1.0340 sec 2025-01-08T02:20:11,402 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(1099): Wait for region servers to report in: status=status unset, state=RUNNING, startTime=1736302811402, completionTime=-1 2025-01-08T02:20:11,402 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.ServerManager(907): Finished waiting on RegionServer count=3; waited=0ms, expected min=3 server(s), max=3 server(s), master is running 2025-01-08T02:20:11,402 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] assignment.AssignmentManager(1747): Joining cluster... 2025-01-08T02:20:11,444 DEBUG [hconnection-0x35e289a5-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:11,446 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33022, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:11,459 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] assignment.AssignmentManager(1759): Number of RegionServers=3 2025-01-08T02:20:11,459 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$RegionInTransitionChore; timeout=60000, timestamp=1736302871459 2025-01-08T02:20:11,459 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.TimeoutExecutorThread(81): ADDED pid=-1, state=WAITING_TIMEOUT; org.apache.hadoop.hbase.master.assignment.AssignmentManager$DeadServerMetricRegionChore; timeout=120000, timestamp=1736302931459 2025-01-08T02:20:11,460 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] assignment.AssignmentManager(1766): Joined the cluster in 57 msec 2025-01-08T02:20:11,479 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] balancer.RegionLocationFinder(172): Locality for region 1588230740 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:20:11,486 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,41063,1736302807767-ClusterStatusChore, period=60000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:11,487 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,41063,1736302807767-BalancerChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:11,487 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,41063,1736302807767-RegionNormalizerChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:11,488 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=CatalogJanitor-c4e5d52b9759:41063, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:11,488 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=HbckChore-, period=3600000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:11,494 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] janitor.CatalogJanitor(179): 2025-01-08T02:20:11,497 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.TableNamespaceManager(92): Namespace table not found. Creating... 2025-01-08T02:20:11,498 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(2440): Client=null/null create 'hbase:namespace', {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T02:20:11,504 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:namespace 2025-01-08T02:20:11,508 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:20:11,509 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:11,511 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:20:11,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741837_1013 (size=358) 2025-01-08T02:20:11,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741837_1013 (size=358) 2025-01-08T02:20:11,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741837_1013 (size=358) 2025-01-08T02:20:11,937 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 4984d49fea614490cee04f171891b535, NAME => 'hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:namespace', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'info', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '10', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:11,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741838_1014 (size=42) 2025-01-08T02:20:11,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741838_1014 (size=42) 2025-01-08T02:20:11,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741838_1014 (size=42) 2025-01-08T02:20:11,952 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:11,952 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1681): Closing 4984d49fea614490cee04f171891b535, disabling compactions & flushes 2025-01-08T02:20:11,952 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:20:11,952 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:20:11,952 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. after waiting 0 ms 2025-01-08T02:20:11,952 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:20:11,952 INFO [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1922): Closed hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:20:11,952 DEBUG [RegionOpenAndInit-hbase:namespace-pool-0 {}] regionserver.HRegion(1635): Region close journal for 4984d49fea614490cee04f171891b535: 2025-01-08T02:20:11,955 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:20:11,963 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535.","families":{"info":[{"qualifier":"regioninfo","vlen":41,"tag":[],"timestamp":"1736302811956"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302811956"}]},"ts":"1736302811956"} 2025-01-08T02:20:11,993 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2025-01-08T02:20:11,996 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:20:12,000 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302811996"}]},"ts":"1736302811996"} 2025-01-08T02:20:12,004 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLING in hbase:meta 2025-01-08T02:20:12,008 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:20:12,010 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:20:12,010 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:20:12,010 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:20:12,011 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:20:12,011 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:20:12,011 INFO [PEWorker-3 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:20:12,011 DEBUG [PEWorker-3 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:20:12,012 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=4984d49fea614490cee04f171891b535, ASSIGN}] 2025-01-08T02:20:12,015 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:namespace, region=4984d49fea614490cee04f171891b535, ASSIGN 2025-01-08T02:20:12,017 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=5, ppid=4, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:namespace, region=4984d49fea614490cee04f171891b535, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:20:12,169 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T02:20:12,169 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=4984d49fea614490cee04f171891b535, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:12,173 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=6, ppid=5, state=RUNNABLE; OpenRegionProcedure 4984d49fea614490cee04f171891b535, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:20:12,329 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:12,341 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(135): Open hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:20:12,341 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7285): Opening region: {ENCODED => 4984d49fea614490cee04f171891b535, NAME => 'hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535.', STARTKEY => '', ENDKEY => ''} 2025-01-08T02:20:12,341 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. service=AccessControlService 2025-01-08T02:20:12,342 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:12,342 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table namespace 4984d49fea614490cee04f171891b535 2025-01-08T02:20:12,342 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(894): Instantiated hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:12,342 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7327): checking encryption for 4984d49fea614490cee04f171891b535 2025-01-08T02:20:12,342 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(7330): checking classloading for 4984d49fea614490cee04f171891b535 2025-01-08T02:20:12,347 INFO [StoreOpener-4984d49fea614490cee04f171891b535-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family info of region 4984d49fea614490cee04f171891b535 2025-01-08T02:20:12,350 INFO [StoreOpener-4984d49fea614490cee04f171891b535-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 4984d49fea614490cee04f171891b535 columnFamilyName info 2025-01-08T02:20:12,350 DEBUG [StoreOpener-4984d49fea614490cee04f171891b535-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:12,351 INFO [StoreOpener-4984d49fea614490cee04f171891b535-1 {}] regionserver.HStore(327): Store=4984d49fea614490cee04f171891b535/info, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:12,353 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535 2025-01-08T02:20:12,353 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535 2025-01-08T02:20:12,358 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1085): writing seq id for 4984d49fea614490cee04f171891b535 2025-01-08T02:20:12,364 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:12,365 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1102): Opened 4984d49fea614490cee04f171891b535; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64824761, jitterRate=-0.03403578698635101}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:12,366 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegion(1001): Region open journal for 4984d49fea614490cee04f171891b535: 2025-01-08T02:20:12,368 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535., pid=6, masterSystemTime=1736302812329 2025-01-08T02:20:12,373 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:20:12,374 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=6}] handler.AssignRegionHandler(164): Opened hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:20:12,375 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=5 updating hbase:meta row=4984d49fea614490cee04f171891b535, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:12,385 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=6, resume processing ppid=5 2025-01-08T02:20:12,387 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=6, ppid=5, state=SUCCESS; OpenRegionProcedure 4984d49fea614490cee04f171891b535, server=c4e5d52b9759,36743,1736302808837 in 206 msec 2025-01-08T02:20:12,391 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=5, resume processing ppid=4 2025-01-08T02:20:12,391 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=5, ppid=4, state=SUCCESS; TransitRegionStateProcedure table=hbase:namespace, region=4984d49fea614490cee04f171891b535, ASSIGN in 373 msec 2025-01-08T02:20:12,393 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:20:12,393 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:namespace","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302812393"}]},"ts":"1736302812393"} 2025-01-08T02:20:12,397 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:namespace, state=ENABLED in hbase:meta 2025-01-08T02:20:12,401 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=4, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:namespace execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:20:12,406 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=4, state=SUCCESS; CreateTableProcedure table=hbase:namespace in 903 msec 2025-01-08T02:20:12,408 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKUtil(113): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/namespace 2025-01-08T02:20:12,410 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:12,410 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:12,410 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:12,415 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/namespace 2025-01-08T02:20:12,415 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:12,465 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=7, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=default 2025-01-08T02:20:12,482 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2025-01-08T02:20:12,487 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=7, state=SUCCESS; CreateNamespaceProcedure, namespace=default in 28 msec 2025-01-08T02:20:12,499 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=8, state=RUNNABLE:CREATE_NAMESPACE_PREPARE; CreateNamespaceProcedure, namespace=hbase 2025-01-08T02:20:12,511 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/namespace 2025-01-08T02:20:12,516 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=8, state=SUCCESS; CreateNamespaceProcedure, namespace=hbase in 16 msec 2025-01-08T02:20:12,527 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/default 2025-01-08T02:20:12,529 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/namespace/hbase 2025-01-08T02:20:12,530 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(1229): Master has completed initialization 3.493sec 2025-01-08T02:20:12,532 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] quotas.MasterQuotaManager(97): Quota support disabled 2025-01-08T02:20:12,534 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] slowlog.SlowLogMasterService(57): Slow/Large requests logging to system table hbase:slowlog is disabled. Quitting. 2025-01-08T02:20:12,535 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] waleventtracker.WALEventTrackerTableCreator(75): wal event tracker requests logging to table REPLICATION.WALEVENTTRACKER is disabled. Quitting. 2025-01-08T02:20:12,536 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.ReplicationSinkTrackerTableCreator(90): replication sink tracker requests logging to table REPLICATION.SINK_TRACKER is disabled. Quitting. 2025-01-08T02:20:12,536 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] zookeeper.ZKWatcher(271): not a secure deployment, proceeding 2025-01-08T02:20:12,537 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,41063,1736302807767-MobFileCleanerChore, period=86400, unit=SECONDS is enabled. 2025-01-08T02:20:12,537 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,41063,1736302807767-MobFileCompactionChore, period=604800, unit=SECONDS is enabled. 2025-01-08T02:20:12,555 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster$4(2404): Client=null/null create 'hbase:acl', {NAME => 'l', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'} 2025-01-08T02:20:12,558 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] procedure2.ProcedureExecutor(1098): Stored pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=hbase:acl 2025-01-08T02:20:12,559 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:20:12,560 DEBUG [PEWorker-2 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:12,560 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] master.MasterRpcServices(713): Client=null/null procedure request for creating table: namespace: "hbase" qualifier: "acl" procId is: 9 2025-01-08T02:20:12,561 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:20:12,595 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T02:20:12,636 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5dda4b49 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@57182bd3 2025-01-08T02:20:12,637 WARN [Time-limited test {}] client.ZKConnectionRegistry(90): ZKConnectionRegistry is deprecated. See https://hbase.apache.org/book.html#client.rpcconnectionregistry 2025-01-08T02:20:12,666 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1fdf1410, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:12,677 DEBUG [Time-limited test {}] nio.NioEventLoop(110): -Dio.netty.noKeySetOptimization: false 2025-01-08T02:20:12,677 DEBUG [Time-limited test {}] nio.NioEventLoop(111): -Dio.netty.selectorAutoRebuildThreshold: 512 2025-01-08T02:20:12,682 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741839_1015 (size=349) 2025-01-08T02:20:12,683 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741839_1015 (size=349) 2025-01-08T02:20:12,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741839_1015 (size=349) 2025-01-08T02:20:12,694 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 8170840180427f0da936e0809aa0d2c2, NAME => 'hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2.', STARTKEY => '', ENDKEY => ''}, tableDescriptor='hbase:acl', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'l', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'NONE', IN_MEMORY => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '8192 B (8KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:12,697 DEBUG [hconnection-0x510316c8-shared-pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:12,698 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T02:20:12,737 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:33036, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1199): Minicluster is up; activeMaster=c4e5d52b9759,41063,1736302807767 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2790): Starting mini mapreduce cluster... 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting test.cache.data to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/test.cache.data in system properties and HBase conf 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.tmp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.tmp.dir in system properties and HBase conf 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting hadoop.log.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir in system properties and HBase conf 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.local.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/mapreduce.cluster.local.dir in system properties and HBase conf 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting mapreduce.cluster.temp.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/mapreduce.cluster.temp.dir in system properties and HBase conf 2025-01-08T02:20:12,743 INFO [Time-limited test {}] hbase.HBaseTestingUtility(811): read short circuit is OFF 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-labels.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.node-labels.fs-store.root-dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.node-attribute.fs-store.root-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.node-attribute.fs-store.root-dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.log-dirs to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.nodemanager.log-dirs in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.active-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.timeline-service.entity-group-fs-store.active-dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.timeline-service.entity-group-fs-store.done-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.timeline-service.entity-group-fs-store.done-dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting yarn.nodemanager.remote-app-log-dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/yarn.nodemanager.remote-app-log-dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.datanode.shared.file.descriptor.paths to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.datanode.shared.file.descriptor.paths in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting nfs.dump.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/nfs.dump.dir in system properties and HBase conf 2025-01-08T02:20:12,744 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting java.io.tmpdir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir in system properties and HBase conf 2025-01-08T02:20:12,745 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.journalnode.edits.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.journalnode.edits.dir in system properties and HBase conf 2025-01-08T02:20:12,745 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting dfs.provided.aliasmap.inmemory.leveldb.dir to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/dfs.provided.aliasmap.inmemory.leveldb.dir in system properties and HBase conf 2025-01-08T02:20:12,745 INFO [Time-limited test {}] hbase.HBaseTestingUtility(824): Setting fs.s3a.committer.staging.tmp.path to /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/fs.s3a.committer.staging.tmp.path in system properties and HBase conf 2025-01-08T02:20:12,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741840_1016 (size=36) 2025-01-08T02:20:12,781 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741840_1016 (size=36) 2025-01-08T02:20:12,781 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741840_1016 (size=36) 2025-01-08T02:20:12,783 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(894): Instantiated hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:12,783 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1681): Closing 8170840180427f0da936e0809aa0d2c2, disabling compactions & flushes 2025-01-08T02:20:12,783 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1703): Closing region hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:20:12,783 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:20:12,784 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. after waiting 1 ms 2025-01-08T02:20:12,784 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:20:12,784 INFO [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1922): Closed hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:20:12,784 DEBUG [RegionOpenAndInit-hbase:acl-pool-0 {}] regionserver.HRegion(1635): Region close journal for 8170840180427f0da936e0809aa0d2c2: 2025-01-08T02:20:12,787 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:20:12,787 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2.","families":{"info":[{"qualifier":"regioninfo","vlen":35,"tag":[],"timestamp":"1736302812787"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302812787"}]},"ts":"1736302812787"} 2025-01-08T02:20:12,795 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 1 regions to meta. 2025-01-08T02:20:12,798 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:20:12,799 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:acl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302812798"}]},"ts":"1736302812798"} 2025-01-08T02:20:12,803 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:acl, state=ENABLING in hbase:meta 2025-01-08T02:20:12,808 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:20:12,813 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:20:12,813 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:20:12,813 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:20:12,813 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:20:12,813 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:20:12,813 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:20:12,813 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:20:12,813 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:acl, region=8170840180427f0da936e0809aa0d2c2, ASSIGN}] 2025-01-08T02:20:12,819 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=hbase:acl, region=8170840180427f0da936e0809aa0d2c2, ASSIGN 2025-01-08T02:20:12,821 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=10, ppid=9, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=hbase:acl, region=8170840180427f0da936e0809aa0d2c2, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:20:12,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741841_1017 (size=592039) 2025-01-08T02:20:12,891 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741841_1017 (size=592039) 2025-01-08T02:20:12,892 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741841_1017 (size=592039) 2025-01-08T02:20:12,898 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T02:20:12,970 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T02:20:12,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T02:20:12,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741842_1018 (size=1663647) 2025-01-08T02:20:12,972 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T02:20:12,972 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=8170840180427f0da936e0809aa0d2c2, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:12,977 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=11, ppid=10, state=RUNNABLE; OpenRegionProcedure 8170840180427f0da936e0809aa0d2c2, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:20:13,157 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:13,157 DEBUG [RSProcedureDispatcher-pool-2 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T02:20:13,185 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:38880, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T02:20:13,199 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T02:20:13,212 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] handler.AssignRegionHandler(135): Open hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:20:13,213 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7285): Opening region: {ENCODED => 8170840180427f0da936e0809aa0d2c2, NAME => 'hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2.', STARTKEY => '', ENDKEY => ''} 2025-01-08T02:20:13,214 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7999): Registered coprocessor service: region=hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. service=AccessControlService 2025-01-08T02:20:13,214 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:13,215 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table acl 8170840180427f0da936e0809aa0d2c2 2025-01-08T02:20:13,215 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(894): Instantiated hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:13,215 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7327): checking encryption for 8170840180427f0da936e0809aa0d2c2 2025-01-08T02:20:13,215 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(7330): checking classloading for 8170840180427f0da936e0809aa0d2c2 2025-01-08T02:20:13,220 INFO [StoreOpener-8170840180427f0da936e0809aa0d2c2-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family l of region 8170840180427f0da936e0809aa0d2c2 2025-01-08T02:20:13,223 INFO [StoreOpener-8170840180427f0da936e0809aa0d2c2-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 8170840180427f0da936e0809aa0d2c2 columnFamilyName l 2025-01-08T02:20:13,223 DEBUG [StoreOpener-8170840180427f0da936e0809aa0d2c2-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:13,226 INFO [StoreOpener-8170840180427f0da936e0809aa0d2c2-1 {}] regionserver.HStore(327): Store=8170840180427f0da936e0809aa0d2c2/l, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:13,227 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2 2025-01-08T02:20:13,228 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2 2025-01-08T02:20:13,232 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1085): writing seq id for 8170840180427f0da936e0809aa0d2c2 2025-01-08T02:20:13,237 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:13,238 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1102): Opened 8170840180427f0da936e0809aa0d2c2; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=59894194, jitterRate=-0.10750696063041687}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:13,241 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegion(1001): Region open journal for 8170840180427f0da936e0809aa0d2c2: 2025-01-08T02:20:13,247 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegionServer(2601): Post open deploy tasks for hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2., pid=11, masterSystemTime=1736302813157 2025-01-08T02:20:13,255 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=10 updating hbase:meta row=8170840180427f0da936e0809aa0d2c2, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:13,255 DEBUG [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] regionserver.HRegionServer(2628): Finished post open deploy task for hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:20:13,256 INFO [RS_OPEN_PRIORITY_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_PRIORITY_REGION, pid=11}] handler.AssignRegionHandler(164): Opened hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:20:13,265 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=11, resume processing ppid=10 2025-01-08T02:20:13,266 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=11, ppid=10, state=SUCCESS; OpenRegionProcedure 8170840180427f0da936e0809aa0d2c2, server=c4e5d52b9759,37031,1736302808764 in 283 msec 2025-01-08T02:20:13,271 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=10, resume processing ppid=9 2025-01-08T02:20:13,271 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=10, ppid=9, state=SUCCESS; TransitRegionStateProcedure table=hbase:acl, region=8170840180427f0da936e0809aa0d2c2, ASSIGN in 453 msec 2025-01-08T02:20:13,274 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:20:13,274 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"hbase:acl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302813274"}]},"ts":"1736302813274"} 2025-01-08T02:20:13,283 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=hbase:acl, state=ENABLED in hbase:meta 2025-01-08T02:20:13,289 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=9, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=hbase:acl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:20:13,293 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=9, state=SUCCESS; CreateTableProcedure table=hbase:acl in 734 msec 2025-01-08T02:20:13,699 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=9 2025-01-08T02:20:13,699 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: hbase:acl, procId: 9 completed 2025-01-08T02:20:13,708 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster {}] master.HMaster(1332): Balancer post startup initialization complete, took 0 seconds 2025-01-08T02:20:13,709 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=RollingUpgradeChore, period=10, unit=SECONDS is enabled. 2025-01-08T02:20:13,710 INFO [master/c4e5d52b9759:0:becomeActiveMaster {}] hbase.ChoreService(168): Chore ScheduledChore name=c4e5d52b9759,41063,1736302807767-OldWALsDirSizeChore, period=300000, unit=MILLISECONDS is enabled. 2025-01-08T02:20:15,055 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:15,193 WARN [Thread-397 {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:15,506 INFO [Thread-397 {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:15,513 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T02:20:15,516 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:15,533 INFO [Thread-397 {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:15,533 INFO [Thread-397 {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:15,533 INFO [Thread-397 {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T02:20:15,536 INFO [Thread-397 {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@4d920d82{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:15,537 INFO [Thread-397 {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@79f1c1ff{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:15,541 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:15,541 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:15,541 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T02:20:15,545 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:15,555 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@520a3669{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:15,556 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@17d5de37{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:15,731 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T02:20:15,732 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices as a root resource class 2025-01-08T02:20:15,732 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T02:20:15,734 INFO [Thread-397 {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T02:20:15,795 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:16,194 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:16,511 INFO [Thread-397 {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:16,538 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@634e8a70{cluster,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-37239-hadoop-yarn-common-3_4_1_jar-_-any-881867231214872516/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/cluster} 2025-01-08T02:20:16,539 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@1dac9f1d{HTTP/1.1, (http/1.1)}{localhost:37239} 2025-01-08T02:20:16,539 INFO [Thread-397 {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@64ae69bd{jobhistory,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-39749-hadoop-yarn-common-3_4_1_jar-_-any-6908857063244387607/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/jobhistory} 2025-01-08T02:20:16,539 INFO [Time-limited test {}] server.Server(415): Started @18045ms 2025-01-08T02:20:16,539 INFO [Thread-397 {}] server.AbstractConnector(333): Started ServerConnector@1fac6a92{HTTP/1.1, (http/1.1)}{localhost:39749} 2025-01-08T02:20:16,540 INFO [Thread-397 {}] server.Server(415): Started @18046ms 2025-01-08T02:20:16,734 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741843_1019 (size=5) 2025-01-08T02:20:16,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741843_1019 (size=5) 2025-01-08T02:20:16,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741843_1019 (size=5) 2025-01-08T02:20:16,831 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:20:16,906 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:meta' 2025-01-08T02:20:16,908 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:namespace' 2025-01-08T02:20:16,910 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'hbase:acl' 2025-01-08T02:20:18,017 WARN [Time-limited test {}] tracker.NMLogAggregationStatusTracker(95): Log Aggregation is disabled.So is the LogAggregationStatusTracker. 2025-01-08T02:20:18,025 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:18,088 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T02:20:18,089 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:18,133 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:18,133 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:18,133 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 600000ms 2025-01-08T02:20:18,147 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:18,161 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6346f6c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:18,161 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@6c4f46cc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:18,320 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2025-01-08T02:20:18,320 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T02:20:18,320 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T02:20:18,320 INFO [Time-limited test {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T02:20:18,332 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:18,363 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:18,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:20:18,432 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase RegionObservers 2025-01-08T02:20:18,433 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_master_table_store 2025-01-08T02:20:18,433 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_master_table_store Metrics about Tables on a single HBase RegionServer 2025-01-08T02:20:18,435 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:20:18,435 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase RegionServerObservers 2025-01-08T02:20:18,435 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2025-01-08T02:20:18,435 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_meta Metrics about Tables on a single HBase RegionServer 2025-01-08T02:20:18,437 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2025-01-08T02:20:18,437 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_namespace Metrics about Tables on a single HBase RegionServer 2025-01-08T02:20:18,441 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_acl 2025-01-08T02:20:18,441 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_hbase_table_acl Metrics about Tables on a single HBase RegionServer 2025-01-08T02:20:18,443 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:20:18,443 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController Metrics about HBase MasterObservers 2025-01-08T02:20:18,443 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T02:20:18,443 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver Metrics about HBase MasterObservers 2025-01-08T02:20:18,446 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T02:20:18,447 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint Metrics about HBase RegionObservers 2025-01-08T02:20:18,591 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:18,637 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@4cc33919{node,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-35457-hadoop-yarn-common-3_4_1_jar-_-any-3502715339511866634/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T02:20:18,639 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@67f55a9f{HTTP/1.1, (http/1.1)}{localhost:35457} 2025-01-08T02:20:18,639 INFO [Time-limited test {}] server.Server(415): Started @20145ms 2025-01-08T02:20:18,939 WARN [Time-limited test {}] tracker.NMLogAggregationStatusTracker(95): Log Aggregation is disabled.So is the LogAggregationStatusTracker. 2025-01-08T02:20:18,943 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:18,960 WARN [Time-limited test {}] servlet.GuiceFilter(102): Multiple Servlet injectors detected. This is a warning indicating that you have more than one GuiceFilter running in your web application. If this is deliberate, you may safely ignore this message. If this is NOT deliberate however, your application may not work as expected. 2025-01-08T02:20:18,961 INFO [Time-limited test {}] server.Server(375): jetty-9.4.53.v20231009; built: 2023-10-09T12:29:09.265Z; git: 27bde00a0b95a1d5bbee0eae7984f891d2d0f8c9; jvm 17.0.11+9 2025-01-08T02:20:18,964 INFO [Time-limited test {}] session.DefaultSessionIdManager(334): DefaultSessionIdManager workerName=node0 2025-01-08T02:20:18,964 INFO [Time-limited test {}] session.DefaultSessionIdManager(339): No SessionScavenger set, using defaults 2025-01-08T02:20:18,964 INFO [Time-limited test {}] session.HouseKeeper(132): node0 Scavenging every 660000ms 2025-01-08T02:20:18,965 WARN [Time-limited test {}] server.AuthenticationFilter(240): Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /home/jenkins/hadoop-http-auth-signature-secret 2025-01-08T02:20:18,965 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@a5b7a6f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,AVAILABLE} 2025-01-08T02:20:18,966 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.s.ServletContextHandler@66f21560{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,AVAILABLE} 2025-01-08T02:20:19,036 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(116): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices as a root resource class 2025-01-08T02:20:19,036 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class 2025-01-08T02:20:19,036 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(113): Registering org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver as a provider class 2025-01-08T02:20:19,037 INFO [Time-limited test {}] application.WebApplicationImpl(815): Initiating Jersey application, version 'Jersey: 1.19.4 05/24/2017 03:20 PM' 2025-01-08T02:20:19,047 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:19,053 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.webapp.GenericExceptionHandler to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:19,147 INFO [Time-limited test {}] container.GuiceComponentProviderFactory(168): Binding org.apache.hadoop.yarn.server.nodemanager.webapp.NMWebServices to GuiceManagedComponentProvider with the scope "Singleton" 2025-01-08T02:20:19,150 INFO [Time-limited test {}] handler.ContextHandler(921): Started o.e.j.w.WebAppContext@3c91a94d{node,/,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/java.io.tmpdir/jetty-localhost-43559-hadoop-yarn-common-3_4_1_jar-_-any-8044330757081492515/webapp/,AVAILABLE}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T02:20:19,151 INFO [Time-limited test {}] server.AbstractConnector(333): Started ServerConnector@402c953e{HTTP/1.1, (http/1.1)}{localhost:43559} 2025-01-08T02:20:19,152 INFO [Time-limited test {}] server.Server(415): Started @20658ms 2025-01-08T02:20:19,185 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2825): Mini mapreduce cluster started 2025-01-08T02:20:19,186 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [30,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:20:19,233 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithTargetName Thread=720, OpenFileDescriptor=781, MaxFileDescriptor=1048576, SystemLoadAverage=883, ProcessCount=18, AvailableMemoryMB=1676 2025-01-08T02:20:19,233 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=720 is superior to 500 2025-01-08T02:20:19,248 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=MasterService, sasl=false 2025-01-08T02:20:19,252 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37610, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=MasterService 2025-01-08T02:20:19,260 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:20:19,266 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=12, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:19,269 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:20:19,270 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithTargetName" procId is: 12 2025-01-08T02:20:19,271 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:20:19,276 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T02:20:19,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741844_1020 (size=442) 2025-01-08T02:20:19,321 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741844_1020 (size=442) 2025-01-08T02:20:19,322 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741844_1020 (size=442) 2025-01-08T02:20:19,327 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 260aa1a42d10faf69822c2b808eb66ed, NAME => 'testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:19,328 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 77eca3710555a2734a4533147eeb6557, NAME => 'testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithTargetName', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:19,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741845_1021 (size=67) 2025-01-08T02:20:19,361 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741845_1021 (size=67) 2025-01-08T02:20:19,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741845_1021 (size=67) 2025-01-08T02:20:19,363 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:19,363 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1681): Closing 260aa1a42d10faf69822c2b808eb66ed, disabling compactions & flushes 2025-01-08T02:20:19,363 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:19,363 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:19,363 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. after waiting 0 ms 2025-01-08T02:20:19,363 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:19,363 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:19,363 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-0 {}] regionserver.HRegion(1635): Region close journal for 260aa1a42d10faf69822c2b808eb66ed: 2025-01-08T02:20:19,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741846_1022 (size=67) 2025-01-08T02:20:19,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741846_1022 (size=67) 2025-01-08T02:20:19,376 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741846_1022 (size=67) 2025-01-08T02:20:19,379 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T02:20:19,381 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:19,381 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1681): Closing 77eca3710555a2734a4533147eeb6557, disabling compactions & flushes 2025-01-08T02:20:19,381 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:19,381 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:19,381 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. after waiting 0 ms 2025-01-08T02:20:19,381 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:19,381 INFO [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:19,381 DEBUG [RegionOpenAndInit-testtb-testExportWithTargetName-pool-1 {}] regionserver.HRegion(1635): Region close journal for 77eca3710555a2734a4533147eeb6557: 2025-01-08T02:20:19,383 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:20:19,384 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.","families":{"info":[{"qualifier":"regioninfo","vlen":66,"tag":[],"timestamp":"1736302819383"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302819383"}]},"ts":"1736302819383"} 2025-01-08T02:20:19,384 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.","families":{"info":[{"qualifier":"regioninfo","vlen":66,"tag":[],"timestamp":"1736302819383"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302819383"}]},"ts":"1736302819383"} 2025-01-08T02:20:19,418 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:20:19,421 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:20:19,421 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302819421"}]},"ts":"1736302819421"} 2025-01-08T02:20:19,424 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=ENABLING in hbase:meta 2025-01-08T02:20:19,429 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:20:19,431 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:20:19,431 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:20:19,431 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:20:19,431 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:20:19,431 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:20:19,431 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:20:19,431 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:20:19,432 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=260aa1a42d10faf69822c2b808eb66ed, ASSIGN}, {pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=77eca3710555a2734a4533147eeb6557, ASSIGN}] 2025-01-08T02:20:19,434 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=77eca3710555a2734a4533147eeb6557, ASSIGN 2025-01-08T02:20:19,434 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=260aa1a42d10faf69822c2b808eb66ed, ASSIGN 2025-01-08T02:20:19,437 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=14, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=77eca3710555a2734a4533147eeb6557, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:20:19,437 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=13, ppid=12, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=260aa1a42d10faf69822c2b808eb66ed, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:20:19,581 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T02:20:19,588 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:20:19,588 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=260aa1a42d10faf69822c2b808eb66ed, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:19,588 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=77eca3710555a2734a4533147eeb6557, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:19,591 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=15, ppid=13, state=RUNNABLE; OpenRegionProcedure 260aa1a42d10faf69822c2b808eb66ed, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:20:19,594 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=16, ppid=14, state=RUNNABLE; OpenRegionProcedure 77eca3710555a2734a4533147eeb6557, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:20:19,746 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:19,748 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:19,760 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(135): Open testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:19,760 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7285): Opening region: {ENCODED => 260aa1a42d10faf69822c2b808eb66ed, NAME => 'testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:20:19,761 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. service=AccessControlService 2025-01-08T02:20:19,761 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:19,762 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithTargetName 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:19,762 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:19,762 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7327): checking encryption for 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:19,762 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(7330): checking classloading for 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:19,788 INFO [StoreOpener-260aa1a42d10faf69822c2b808eb66ed-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:19,801 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] handler.AssignRegionHandler(135): Open testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:19,802 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7285): Opening region: {ENCODED => 77eca3710555a2734a4533147eeb6557, NAME => 'testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:20:19,803 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. service=AccessControlService 2025-01-08T02:20:19,803 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:19,803 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithTargetName 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:19,804 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(894): Instantiated testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:19,804 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7327): checking encryption for 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:19,804 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(7330): checking classloading for 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:19,811 INFO [StoreOpener-260aa1a42d10faf69822c2b808eb66ed-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 260aa1a42d10faf69822c2b808eb66ed columnFamilyName cf 2025-01-08T02:20:19,817 INFO [StoreOpener-77eca3710555a2734a4533147eeb6557-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:19,824 DEBUG [StoreOpener-260aa1a42d10faf69822c2b808eb66ed-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:19,829 INFO [StoreOpener-260aa1a42d10faf69822c2b808eb66ed-1 {}] regionserver.HStore(327): Store=260aa1a42d10faf69822c2b808eb66ed/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:19,832 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:19,833 INFO [StoreOpener-77eca3710555a2734a4533147eeb6557-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 77eca3710555a2734a4533147eeb6557 columnFamilyName cf 2025-01-08T02:20:19,835 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:19,835 DEBUG [StoreOpener-77eca3710555a2734a4533147eeb6557-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:19,842 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1085): writing seq id for 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:19,845 INFO [StoreOpener-77eca3710555a2734a4533147eeb6557-1 {}] regionserver.HStore(327): Store=77eca3710555a2734a4533147eeb6557/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:19,848 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:19,849 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:19,855 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1085): writing seq id for 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:19,868 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:19,870 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:19,870 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1102): Opened 260aa1a42d10faf69822c2b808eb66ed; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68507867, jitterRate=0.020846769213676453}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:19,871 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1102): Opened 77eca3710555a2734a4533147eeb6557; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64966485, jitterRate=-0.03192393481731415}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:19,873 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegion(1001): Region open journal for 77eca3710555a2734a4533147eeb6557: 2025-01-08T02:20:19,873 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegion(1001): Region open journal for 260aa1a42d10faf69822c2b808eb66ed: 2025-01-08T02:20:19,875 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed., pid=15, masterSystemTime=1736302819746 2025-01-08T02:20:19,877 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557., pid=16, masterSystemTime=1736302819748 2025-01-08T02:20:19,880 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:19,881 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=15}] handler.AssignRegionHandler(164): Opened testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:19,881 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=13 updating hbase:meta row=260aa1a42d10faf69822c2b808eb66ed, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:19,884 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=14 updating hbase:meta row=77eca3710555a2734a4533147eeb6557, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:19,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T02:20:19,885 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:19,885 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=16}] handler.AssignRegionHandler(164): Opened testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:19,891 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=15, resume processing ppid=13 2025-01-08T02:20:19,892 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=15, ppid=13, state=SUCCESS; OpenRegionProcedure 260aa1a42d10faf69822c2b808eb66ed, server=c4e5d52b9759,37031,1736302808764 in 294 msec 2025-01-08T02:20:19,896 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=16, resume processing ppid=14 2025-01-08T02:20:19,896 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=16, ppid=14, state=SUCCESS; OpenRegionProcedure 77eca3710555a2734a4533147eeb6557, server=c4e5d52b9759,36743,1736302808837 in 295 msec 2025-01-08T02:20:19,898 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=13, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=260aa1a42d10faf69822c2b808eb66ed, ASSIGN in 460 msec 2025-01-08T02:20:19,900 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=14, resume processing ppid=12 2025-01-08T02:20:19,900 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=14, ppid=12, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=77eca3710555a2734a4533147eeb6557, ASSIGN in 464 msec 2025-01-08T02:20:19,902 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:20:19,902 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302819902"}]},"ts":"1736302819902"} 2025-01-08T02:20:19,905 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=ENABLED in hbase:meta 2025-01-08T02:20:19,912 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=12, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithTargetName execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:20:19,916 DEBUG [PEWorker-1 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithTargetName jenkins: RWXCA 2025-01-08T02:20:19,921 DEBUG [PEWorker-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:19,925 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60156, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:19,945 DEBUG [hconnection-0x63535853-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:19,947 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56084, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=ClientService 2025-01-08T02:20:19,955 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T02:20:19,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:19,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:19,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:19,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:19,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:19,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeCreated, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:19,995 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:19,995 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:20:20,009 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=12, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithTargetName in 743 msec 2025-01-08T02:20:20,030 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:20,031 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:20,041 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:20,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF\x0AE\x0A\x07jenkins\x12:\x08\x03"6\x0A*\x0A\x07default\x12\x1Ftesttb-testExportWithTargetName \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:20,388 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=12 2025-01-08T02:20:20,389 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithTargetName, procId: 12 completed 2025-01-08T02:20:20,393 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithTargetName 2025-01-08T02:20:20,394 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:20,395 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:20:20,417 DEBUG [hconnection-0x510316c8-shared-pool-4 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:20,420 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60172, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:20,435 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T02:20:20,435 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302820435 (current time:1736302820435). 2025-01-08T02:20:20,435 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:20:20,435 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithTargetName VERSION not specified, setting to 2 2025-01-08T02:20:20,436 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:20:20,437 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x66e3815f to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@134e944e 2025-01-08T02:20:20,448 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@274483a2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:20,453 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:20,454 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56096, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:20,457 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x66e3815f to 127.0.0.1:61593 2025-01-08T02:20:20,458 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:20,461 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4f9e654c to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5cce9f1c 2025-01-08T02:20:20,473 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1aea8100, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:20,476 DEBUG [hconnection-0x6beb2a46-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:20,478 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56106, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:20,481 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:20,482 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60184, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:20,483 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4f9e654c to 127.0.0.1:61593 2025-01-08T02:20:20,483 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:20,484 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T02:20:20,498 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:20:20,506 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=17, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T02:20:20,507 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 17 2025-01-08T02:20:20,509 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T02:20:20,513 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:20:20,518 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:20:20,533 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:20:20,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741847_1023 (size=167) 2025-01-08T02:20:20,548 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741847_1023 (size=167) 2025-01-08T02:20:20,549 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741847_1023 (size=167) 2025-01-08T02:20:20,550 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:20:20,553 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed}, {pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557}] 2025-01-08T02:20:20,557 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:20,558 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:20,611 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T02:20:20,713 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:20,713 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:20,715 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=18 2025-01-08T02:20:20,715 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=19 2025-01-08T02:20:20,720 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:20,720 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:20,722 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.HRegion(2538): Flush status journal for 260aa1a42d10faf69822c2b808eb66ed: 2025-01-08T02:20:20,722 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. for emptySnaptb0-testExportWithTargetName completed. 2025-01-08T02:20:20,723 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.' region-info for snapshot=emptySnaptb0-testExportWithTargetName 2025-01-08T02:20:20,723 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.HRegion(2538): Flush status journal for 77eca3710555a2734a4533147eeb6557: 2025-01-08T02:20:20,723 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. for emptySnaptb0-testExportWithTargetName completed. 2025-01-08T02:20:20,723 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.' region-info for snapshot=emptySnaptb0-testExportWithTargetName 2025-01-08T02:20:20,726 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:20,726 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:20,729 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:20:20,729 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:20:20,748 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741848_1024 (size=70) 2025-01-08T02:20:20,749 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741848_1024 (size=70) 2025-01-08T02:20:20,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741848_1024 (size=70) 2025-01-08T02:20:20,752 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:20,754 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=18}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=18 2025-01-08T02:20:20,757 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741849_1025 (size=70) 2025-01-08T02:20:20,757 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=18 2025-01-08T02:20:20,757 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithTargetName on region 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:20,757 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=18, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:20,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741849_1025 (size=70) 2025-01-08T02:20:20,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741849_1025 (size=70) 2025-01-08T02:20:20,759 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:20,760 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=19}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=19 2025-01-08T02:20:20,760 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=19 2025-01-08T02:20:20,761 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithTargetName on region 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:20,761 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=19, ppid=17, state=RUNNABLE; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:20,766 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=18, ppid=17, state=SUCCESS; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed in 206 msec 2025-01-08T02:20:20,768 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=19, resume processing ppid=17 2025-01-08T02:20:20,769 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=19, ppid=17, state=SUCCESS; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557 in 211 msec 2025-01-08T02:20:20,769 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:20:20,771 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:20:20,774 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:20:20,774 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:20:20,774 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:20,775 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:20:20,789 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741850_1026 (size=62) 2025-01-08T02:20:20,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741850_1026 (size=62) 2025-01-08T02:20:20,791 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741850_1026 (size=62) 2025-01-08T02:20:20,793 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:20:20,793 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithTargetName 2025-01-08T02:20:20,796 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithTargetName 2025-01-08T02:20:20,813 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T02:20:20,850 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741851_1027 (size=649) 2025-01-08T02:20:20,850 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741851_1027 (size=649) 2025-01-08T02:20:20,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741851_1027 (size=649) 2025-01-08T02:20:20,861 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:20:20,878 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:20:20,879 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithTargetName to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testExportWithTargetName 2025-01-08T02:20:20,883 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=17, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:20:20,883 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 17 2025-01-08T02:20:20,886 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=17, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=17, snapshot={ ss=emptySnaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } in 381 msec 2025-01-08T02:20:21,116 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=17 2025-01-08T02:20:21,116 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithTargetName, procId: 17 completed 2025-01-08T02:20:21,134 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:20:21,145 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:20:21,154 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithTargetName 2025-01-08T02:20:21,154 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:21,155 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:20:21,182 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T02:20:21,182 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302821182 (current time:1736302821182). 2025-01-08T02:20:21,182 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:20:21,183 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithTargetName VERSION not specified, setting to 2 2025-01-08T02:20:21,183 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:20:21,185 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5dcebeac to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@56d7847 2025-01-08T02:20:21,221 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@44cb4011, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:21,224 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:21,226 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56110, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:21,228 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5dcebeac to 127.0.0.1:61593 2025-01-08T02:20:21,228 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:21,231 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x577168a4 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6bd260f9 2025-01-08T02:20:21,261 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@60714cb8, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:21,286 DEBUG [hconnection-0xa2dcc45-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:21,288 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56118, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:21,291 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:21,293 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:60198, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:21,294 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x577168a4 to 127.0.0.1:61593 2025-01-08T02:20:21,295 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:21,295 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithTargetName], kv [jenkins: RWXCA] 2025-01-08T02:20:21,296 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:20:21,298 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=20, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } 2025-01-08T02:20:21,298 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 20 2025-01-08T02:20:21,300 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:20:21,302 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:20:21,303 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T02:20:21,306 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:20:21,388 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741852_1028 (size=162) 2025-01-08T02:20:21,388 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741852_1028 (size=162) 2025-01-08T02:20:21,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741852_1028 (size=162) 2025-01-08T02:20:21,397 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:20:21,397 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed}, {pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557}] 2025-01-08T02:20:21,399 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:21,400 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:21,410 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T02:20:21,553 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:21,553 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:21,554 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=21 2025-01-08T02:20:21,556 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=22 2025-01-08T02:20:21,557 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:21,557 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(2837): Flushing 260aa1a42d10faf69822c2b808eb66ed 1/1 column families, dataSize=132 B heapSize=544 B 2025-01-08T02:20:21,561 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:21,561 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(2837): Flushing 77eca3710555a2734a4533147eeb6557 1/1 column families, dataSize=3.13 KB heapSize=7 KB 2025-01-08T02:20:21,612 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T02:20:21,705 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108498e4d60a68145b18d3384bad66c27e0_260aa1a42d10faf69822c2b808eb66ed is 71, key is 06a7fac9c8dadaa46921ec836bcec6fc/cf:q/1736302821134/Put/seqid=0 2025-01-08T02:20:21,714 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108f6f537b1374b458a9364196e607979dd_77eca3710555a2734a4533147eeb6557 is 71, key is 1ae498863d06deba200811f5986044d3/cf:q/1736302821145/Put/seqid=0 2025-01-08T02:20:21,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741853_1029 (size=8242) 2025-01-08T02:20:21,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741853_1029 (size=8242) 2025-01-08T02:20:21,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741853_1029 (size=8242) 2025-01-08T02:20:21,822 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:21,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741854_1030 (size=5032) 2025-01-08T02:20:21,847 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741854_1030 (size=5032) 2025-01-08T02:20:21,847 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741854_1030 (size=5032) 2025-01-08T02:20:21,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:21,915 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T02:20:21,956 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108498e4d60a68145b18d3384bad66c27e0_260aa1a42d10faf69822c2b808eb66ed to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e20250108498e4d60a68145b18d3384bad66c27e0_260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:21,957 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108f6f537b1374b458a9364196e607979dd_77eca3710555a2734a4533147eeb6557 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6f537b1374b458a9364196e607979dd_77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:21,959 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/.tmp/cf/96698aec059d4e468c54479121710b2a, store: [table=testtb-testExportWithTargetName family=cf region=260aa1a42d10faf69822c2b808eb66ed] 2025-01-08T02:20:21,972 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/.tmp/cf/9dfd847383954939bb9be9923f8f52b4, store: [table=testtb-testExportWithTargetName family=cf region=77eca3710555a2734a4533147eeb6557] 2025-01-08T02:20:21,984 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/.tmp/cf/9dfd847383954939bb9be9923f8f52b4 is 208, key is 14e4c6f7a1585525f9731255649ebf914/cf:q/1736302821145/Put/seqid=0 2025-01-08T02:20:21,986 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/.tmp/cf/96698aec059d4e468c54479121710b2a is 208, key is 0c92402f9361d2db40a27946cfbedc183/cf:q/1736302821134/Put/seqid=0 2025-01-08T02:20:22,064 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741855_1031 (size=15155) 2025-01-08T02:20:22,066 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741855_1031 (size=15155) 2025-01-08T02:20:22,067 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741855_1031 (size=15155) 2025-01-08T02:20:22,068 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/.tmp/cf/9dfd847383954939bb9be9923f8f52b4 2025-01-08T02:20:22,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741856_1032 (size=5706) 2025-01-08T02:20:22,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741856_1032 (size=5706) 2025-01-08T02:20:22,084 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=132, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/.tmp/cf/96698aec059d4e468c54479121710b2a 2025-01-08T02:20:22,084 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741856_1032 (size=5706) 2025-01-08T02:20:22,097 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/.tmp/cf/9dfd847383954939bb9be9923f8f52b4 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/cf/9dfd847383954939bb9be9923f8f52b4 2025-01-08T02:20:22,106 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/.tmp/cf/96698aec059d4e468c54479121710b2a as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/cf/96698aec059d4e468c54479121710b2a 2025-01-08T02:20:22,114 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/cf/9dfd847383954939bb9be9923f8f52b4, entries=48, sequenceid=6, filesize=14.8 K 2025-01-08T02:20:22,117 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/cf/96698aec059d4e468c54479121710b2a, entries=2, sequenceid=6, filesize=5.6 K 2025-01-08T02:20:22,118 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(3040): Finished flush of dataSize ~132 B/132, heapSize ~528 B/528, currentSize=0 B/0 for 260aa1a42d10faf69822c2b808eb66ed in 561ms, sequenceid=6, compaction requested=false 2025-01-08T02:20:22,119 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(3040): Finished flush of dataSize ~3.13 KB/3204, heapSize ~6.98 KB/7152, currentSize=0 B/0 for 77eca3710555a2734a4533147eeb6557 in 557ms, sequenceid=6, compaction requested=false 2025-01-08T02:20:22,119 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithTargetName' 2025-01-08T02:20:22,119 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithTargetName' 2025-01-08T02:20:22,120 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.HRegion(2538): Flush status journal for 77eca3710555a2734a4533147eeb6557: 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.HRegion(2538): Flush status journal for 260aa1a42d10faf69822c2b808eb66ed: 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. for snaptb0-testExportWithTargetName completed. 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. for snaptb0-testExportWithTargetName completed. 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.' region-info for snapshot=snaptb0-testExportWithTargetName 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/cf/96698aec059d4e468c54479121710b2a] hfiles 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/cf/96698aec059d4e468c54479121710b2a for snapshot=snaptb0-testExportWithTargetName 2025-01-08T02:20:22,121 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.' region-info for snapshot=snaptb0-testExportWithTargetName 2025-01-08T02:20:22,122 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:22,122 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/cf/9dfd847383954939bb9be9923f8f52b4] hfiles 2025-01-08T02:20:22,122 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/cf/9dfd847383954939bb9be9923f8f52b4 for snapshot=snaptb0-testExportWithTargetName 2025-01-08T02:20:22,142 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741857_1033 (size=109) 2025-01-08T02:20:22,142 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741857_1033 (size=109) 2025-01-08T02:20:22,143 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741857_1033 (size=109) 2025-01-08T02:20:22,143 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:22,143 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=21}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=21 2025-01-08T02:20:22,144 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=21 2025-01-08T02:20:22,144 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithTargetName on region 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:22,144 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=21, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:22,149 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=21, ppid=20, state=SUCCESS; SnapshotRegionProcedure 260aa1a42d10faf69822c2b808eb66ed in 749 msec 2025-01-08T02:20:22,151 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741858_1034 (size=109) 2025-01-08T02:20:22,151 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741858_1034 (size=109) 2025-01-08T02:20:22,152 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741858_1034 (size=109) 2025-01-08T02:20:22,153 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:22,153 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=22}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=22 2025-01-08T02:20:22,154 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=22 2025-01-08T02:20:22,154 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithTargetName on region 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:22,154 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=22, ppid=20, state=RUNNABLE; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:22,162 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=22, resume processing ppid=20 2025-01-08T02:20:22,162 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=22, ppid=20, state=SUCCESS; SnapshotRegionProcedure 77eca3710555a2734a4533147eeb6557 in 758 msec 2025-01-08T02:20:22,162 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:20:22,163 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:20:22,166 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:20:22,166 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:20:22,167 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:22,169 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6f537b1374b458a9364196e607979dd_77eca3710555a2734a4533147eeb6557, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e20250108498e4d60a68145b18d3384bad66c27e0_260aa1a42d10faf69822c2b808eb66ed] hfiles 2025-01-08T02:20:22,169 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6f537b1374b458a9364196e607979dd_77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:22,170 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e20250108498e4d60a68145b18d3384bad66c27e0_260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:22,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741859_1035 (size=293) 2025-01-08T02:20:22,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741859_1035 (size=293) 2025-01-08T02:20:22,202 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741859_1035 (size=293) 2025-01-08T02:20:22,204 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:20:22,204 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithTargetName 2025-01-08T02:20:22,205 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportWithTargetName 2025-01-08T02:20:22,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741860_1036 (size=959) 2025-01-08T02:20:22,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741860_1036 (size=959) 2025-01-08T02:20:22,240 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741860_1036 (size=959) 2025-01-08T02:20:22,270 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:20:22,319 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:20:22,320 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportWithTargetName to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithTargetName 2025-01-08T02:20:22,323 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=20, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:20:22,323 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 }, snapshot procedure id = 20 2025-01-08T02:20:22,327 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=20, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=20, snapshot={ ss=snaptb0-testExportWithTargetName table=testtb-testExportWithTargetName type=FLUSH ttl=0 } in 1.0270 sec 2025-01-08T02:20:22,421 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=20 2025-01-08T02:20:22,423 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithTargetName, procId: 20 completed 2025-01-08T02:20:22,424 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423 2025-01-08T02:20:22,424 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:22,584 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:22,584 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423/.hbase-snapshot/.tmp/testExportWithTargetName 2025-01-08T02:20:22,595 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:20:22,639 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithTargetName to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423/.hbase-snapshot/.tmp/testExportWithTargetName 2025-01-08T02:20:22,753 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741861_1037 (size=162) 2025-01-08T02:20:22,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741861_1037 (size=162) 2025-01-08T02:20:22,754 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741861_1037 (size=162) 2025-01-08T02:20:22,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741862_1038 (size=959) 2025-01-08T02:20:22,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741862_1038 (size=959) 2025-01-08T02:20:22,791 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741862_1038 (size=959) 2025-01-08T02:20:23,171 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741863_1039 (size=154) 2025-01-08T02:20:23,172 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741863_1039 (size=154) 2025-01-08T02:20:23,172 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741863_1039 (size=154) 2025-01-08T02:20:23,180 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:23,180 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:23,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:23,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,364 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-7974865213998038757.jar 2025-01-08T02:20:24,365 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,365 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,468 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-17521981156196860150.jar 2025-01-08T02:20:24,469 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,469 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,470 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,470 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,471 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,471 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:24,472 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:20:24,472 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:20:24,473 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:20:24,473 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:20:24,474 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:20:24,474 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:20:24,475 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:20:24,475 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:20:24,476 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:20:24,476 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:20:24,477 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:20:24,477 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:20:24,480 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:24,481 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:24,481 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:20:24,482 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:24,482 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:24,483 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:20:24,483 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:20:24,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741864_1040 (size=127628) 2025-01-08T02:20:24,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741864_1040 (size=127628) 2025-01-08T02:20:24,663 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741864_1040 (size=127628) 2025-01-08T02:20:24,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T02:20:24,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T02:20:24,706 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741865_1041 (size=2172137) 2025-01-08T02:20:24,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741866_1042 (size=213228) 2025-01-08T02:20:24,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741866_1042 (size=213228) 2025-01-08T02:20:24,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741866_1042 (size=213228) 2025-01-08T02:20:24,767 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741867_1043 (size=6350715) 2025-01-08T02:20:24,767 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741867_1043 (size=6350715) 2025-01-08T02:20:24,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741867_1043 (size=6350715) 2025-01-08T02:20:24,811 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741868_1044 (size=1877034) 2025-01-08T02:20:24,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741868_1044 (size=1877034) 2025-01-08T02:20:24,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741868_1044 (size=1877034) 2025-01-08T02:20:24,834 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741869_1045 (size=533455) 2025-01-08T02:20:24,834 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741869_1045 (size=533455) 2025-01-08T02:20:24,835 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741869_1045 (size=533455) 2025-01-08T02:20:24,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741870_1046 (size=7280644) 2025-01-08T02:20:24,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741870_1046 (size=7280644) 2025-01-08T02:20:24,902 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741870_1046 (size=7280644) 2025-01-08T02:20:24,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741871_1047 (size=4188619) 2025-01-08T02:20:24,975 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741871_1047 (size=4188619) 2025-01-08T02:20:24,977 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741871_1047 (size=4188619) 2025-01-08T02:20:25,012 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741872_1048 (size=20406) 2025-01-08T02:20:25,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741872_1048 (size=20406) 2025-01-08T02:20:25,016 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741872_1048 (size=20406) 2025-01-08T02:20:25,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741873_1049 (size=75495) 2025-01-08T02:20:25,060 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741873_1049 (size=75495) 2025-01-08T02:20:25,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741873_1049 (size=75495) 2025-01-08T02:20:25,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741874_1050 (size=45609) 2025-01-08T02:20:25,116 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741874_1050 (size=45609) 2025-01-08T02:20:25,117 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741874_1050 (size=45609) 2025-01-08T02:20:25,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741875_1051 (size=110084) 2025-01-08T02:20:25,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741875_1051 (size=110084) 2025-01-08T02:20:25,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741875_1051 (size=110084) 2025-01-08T02:20:25,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741876_1052 (size=1323991) 2025-01-08T02:20:25,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741876_1052 (size=1323991) 2025-01-08T02:20:25,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741876_1052 (size=1323991) 2025-01-08T02:20:25,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741877_1053 (size=23076) 2025-01-08T02:20:25,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741877_1053 (size=23076) 2025-01-08T02:20:25,230 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741877_1053 (size=23076) 2025-01-08T02:20:25,256 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741878_1054 (size=126803) 2025-01-08T02:20:25,256 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741878_1054 (size=126803) 2025-01-08T02:20:25,257 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741878_1054 (size=126803) 2025-01-08T02:20:25,265 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:20:25,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741879_1055 (size=322274) 2025-01-08T02:20:25,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741879_1055 (size=322274) 2025-01-08T02:20:25,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741879_1055 (size=322274) 2025-01-08T02:20:25,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741880_1056 (size=1832290) 2025-01-08T02:20:25,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741880_1056 (size=1832290) 2025-01-08T02:20:25,347 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741880_1056 (size=1832290) 2025-01-08T02:20:25,383 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741881_1057 (size=30081) 2025-01-08T02:20:25,384 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741881_1057 (size=30081) 2025-01-08T02:20:25,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741881_1057 (size=30081) 2025-01-08T02:20:25,406 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741882_1058 (size=53616) 2025-01-08T02:20:25,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741882_1058 (size=53616) 2025-01-08T02:20:25,407 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741882_1058 (size=53616) 2025-01-08T02:20:25,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741883_1059 (size=29229) 2025-01-08T02:20:25,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741883_1059 (size=29229) 2025-01-08T02:20:25,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741883_1059 (size=29229) 2025-01-08T02:20:25,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741884_1060 (size=169089) 2025-01-08T02:20:25,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741884_1060 (size=169089) 2025-01-08T02:20:25,445 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741884_1060 (size=169089) 2025-01-08T02:20:25,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741885_1061 (size=451756) 2025-01-08T02:20:25,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741885_1061 (size=451756) 2025-01-08T02:20:25,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741885_1061 (size=451756) 2025-01-08T02:20:25,957 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T02:20:25,958 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T02:20:25,969 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741886_1062 (size=5175431) 2025-01-08T02:20:25,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741887_1063 (size=136454) 2025-01-08T02:20:25,996 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741887_1063 (size=136454) 2025-01-08T02:20:25,996 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741887_1063 (size=136454) 2025-01-08T02:20:26,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741888_1064 (size=907469) 2025-01-08T02:20:26,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741888_1064 (size=907469) 2025-01-08T02:20:26,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741888_1064 (size=907469) 2025-01-08T02:20:26,126 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741889_1065 (size=3317408) 2025-01-08T02:20:26,126 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741889_1065 (size=3317408) 2025-01-08T02:20:26,131 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741889_1065 (size=3317408) 2025-01-08T02:20:26,184 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741890_1066 (size=503880) 2025-01-08T02:20:26,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741890_1066 (size=503880) 2025-01-08T02:20:26,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741890_1066 (size=503880) 2025-01-08T02:20:26,241 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T02:20:26,242 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T02:20:26,243 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741891_1067 (size=4695811) 2025-01-08T02:20:26,249 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:20:26,258 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithTargetName' hfile list 2025-01-08T02:20:26,265 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.3 K 2025-01-08T02:20:26,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741892_1068 (size=722) 2025-01-08T02:20:26,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741892_1068 (size=722) 2025-01-08T02:20:26,292 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741892_1068 (size=722) 2025-01-08T02:20:26,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741893_1069 (size=15) 2025-01-08T02:20:26,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741893_1069 (size=15) 2025-01-08T02:20:26,334 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741893_1069 (size=15) 2025-01-08T02:20:26,514 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741894_1070 (size=304884) 2025-01-08T02:20:26,517 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741894_1070 (size=304884) 2025-01-08T02:20:26,526 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741894_1070 (size=304884) 2025-01-08T02:20:26,921 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:20:26,921 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:20:27,291 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0001_000001 (auth:SIMPLE) from 127.0.0.1:34020 2025-01-08T02:20:28,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName 2025-01-08T02:20:28,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName Metrics about Tables on a single HBase RegionServer 2025-01-08T02:20:34,688 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0001_000001 (auth:SIMPLE) from 127.0.0.1:60672 2025-01-08T02:20:35,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741895_1071 (size=350558) 2025-01-08T02:20:35,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741895_1071 (size=350558) 2025-01-08T02:20:35,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741895_1071 (size=350558) 2025-01-08T02:20:36,388 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:20:37,074 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0001_000001 (auth:SIMPLE) from 127.0.0.1:40224 2025-01-08T02:20:40,348 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T02:20:40,350 INFO [RS-EventLoopGroup-1-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:58388, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.1 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T02:20:41,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741896_1072 (size=15155) 2025-01-08T02:20:41,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741896_1072 (size=15155) 2025-01-08T02:20:41,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741896_1072 (size=15155) 2025-01-08T02:20:41,632 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741897_1073 (size=8242) 2025-01-08T02:20:41,633 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741897_1073 (size=8242) 2025-01-08T02:20:41,633 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741897_1073 (size=8242) 2025-01-08T02:20:41,777 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741898_1074 (size=5706) 2025-01-08T02:20:41,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741898_1074 (size=5706) 2025-01-08T02:20:41,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741898_1074 (size=5706) 2025-01-08T02:20:41,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741899_1075 (size=5032) 2025-01-08T02:20:41,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741899_1075 (size=5032) 2025-01-08T02:20:41,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741899_1075 (size=5032) 2025-01-08T02:20:42,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741900_1076 (size=17461) 2025-01-08T02:20:42,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741900_1076 (size=17461) 2025-01-08T02:20:42,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741900_1076 (size=17461) 2025-01-08T02:20:42,297 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0001/container_1736302816589_0001_01_000002/launch_container.sh] 2025-01-08T02:20:42,297 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0001/container_1736302816589_0001_01_000002/container_tokens] 2025-01-08T02:20:42,298 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0001/container_1736302816589_0001_01_000002/sysfs] 2025-01-08T02:20:42,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741901_1077 (size=464) 2025-01-08T02:20:42,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741901_1077 (size=464) 2025-01-08T02:20:42,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741901_1077 (size=464) 2025-01-08T02:20:42,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741902_1078 (size=17461) 2025-01-08T02:20:42,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741902_1078 (size=17461) 2025-01-08T02:20:42,717 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741902_1078 (size=17461) 2025-01-08T02:20:42,758 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741903_1079 (size=350558) 2025-01-08T02:20:42,762 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741903_1079 (size=350558) 2025-01-08T02:20:42,764 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741903_1079 (size=350558) 2025-01-08T02:20:42,783 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0001_000001 (auth:SIMPLE) from 127.0.0.1:40232 2025-01-08T02:20:43,335 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T02:20:43,336 INFO [RS-EventLoopGroup-1-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:58394, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.0 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T02:20:43,345 DEBUG [BootstrapNodeManager {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=RegionServerStatusService, sasl=false 2025-01-08T02:20:43,346 INFO [RS-EventLoopGroup-1-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:58402, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins.hfs.2 (auth:SIMPLE), service=RegionServerStatusService 2025-01-08T02:20:44,324 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:20:44,325 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:20:44,337 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: testExportWithTargetName 2025-01-08T02:20:44,337 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:20:44,338 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:20:44,338 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithTargetName at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithTargetName 2025-01-08T02:20:44,339 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithTargetName/.snapshotinfo 2025-01-08T02:20:44,339 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithTargetName/data.manifest 2025-01-08T02:20:44,339 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423/.hbase-snapshot/testExportWithTargetName at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423/.hbase-snapshot/testExportWithTargetName 2025-01-08T02:20:44,340 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423/.hbase-snapshot/testExportWithTargetName/.snapshotinfo 2025-01-08T02:20:44,340 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302822423/.hbase-snapshot/testExportWithTargetName/data.manifest 2025-01-08T02:20:44,361 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithTargetName 2025-01-08T02:20:44,366 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportWithTargetName 2025-01-08T02:20:44,372 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=23, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:44,377 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T02:20:44,377 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302844377"}]},"ts":"1736302844377"} 2025-01-08T02:20:44,380 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=DISABLING in hbase:meta 2025-01-08T02:20:44,383 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithTargetName to state=DISABLING 2025-01-08T02:20:44,386 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=24, ppid=23, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithTargetName}] 2025-01-08T02:20:44,392 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=25, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=260aa1a42d10faf69822c2b808eb66ed, UNASSIGN}, {pid=26, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=77eca3710555a2734a4533147eeb6557, UNASSIGN}] 2025-01-08T02:20:44,393 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=26, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=77eca3710555a2734a4533147eeb6557, UNASSIGN 2025-01-08T02:20:44,394 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=25, ppid=24, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=260aa1a42d10faf69822c2b808eb66ed, UNASSIGN 2025-01-08T02:20:44,395 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=25 updating hbase:meta row=260aa1a42d10faf69822c2b808eb66ed, regionState=CLOSING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:44,395 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=26 updating hbase:meta row=77eca3710555a2734a4533147eeb6557, regionState=CLOSING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:44,398 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:20:44,398 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=27, ppid=26, state=RUNNABLE; CloseRegionProcedure 77eca3710555a2734a4533147eeb6557, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:20:44,399 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:20:44,403 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=28, ppid=25, state=RUNNABLE; CloseRegionProcedure 260aa1a42d10faf69822c2b808eb66ed, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:20:44,479 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T02:20:44,557 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:44,557 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:44,560 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(124): Close 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:44,560 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(124): Close 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:44,560 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:20:44,560 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1681): Closing 77eca3710555a2734a4533147eeb6557, disabling compactions & flushes 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1681): Closing 260aa1a42d10faf69822c2b808eb66ed, disabling compactions & flushes 2025-01-08T02:20:44,561 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:44,561 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1703): Closing region testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. after waiting 0 ms 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. after waiting 0 ms 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:44,561 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:44,568 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:20:44,569 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:20:44,572 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:20:44,572 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:20:44,573 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557. 2025-01-08T02:20:44,573 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1922): Closed testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed. 2025-01-08T02:20:44,573 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] regionserver.HRegion(1635): Region close journal for 77eca3710555a2734a4533147eeb6557: 2025-01-08T02:20:44,573 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] regionserver.HRegion(1635): Region close journal for 260aa1a42d10faf69822c2b808eb66ed: 2025-01-08T02:20:44,575 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=28}] handler.UnassignRegionHandler(170): Closed 260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:44,576 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=25 updating hbase:meta row=260aa1a42d10faf69822c2b808eb66ed, regionState=CLOSED 2025-01-08T02:20:44,576 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=27}] handler.UnassignRegionHandler(170): Closed 77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:44,577 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=26 updating hbase:meta row=77eca3710555a2734a4533147eeb6557, regionState=CLOSED 2025-01-08T02:20:44,581 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=28, resume processing ppid=25 2025-01-08T02:20:44,581 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=28, ppid=25, state=SUCCESS; CloseRegionProcedure 260aa1a42d10faf69822c2b808eb66ed, server=c4e5d52b9759,37031,1736302808764 in 179 msec 2025-01-08T02:20:44,583 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=27, resume processing ppid=26 2025-01-08T02:20:44,583 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=27, ppid=26, state=SUCCESS; CloseRegionProcedure 77eca3710555a2734a4533147eeb6557, server=c4e5d52b9759,36743,1736302808837 in 181 msec 2025-01-08T02:20:44,584 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=25, ppid=24, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=260aa1a42d10faf69822c2b808eb66ed, UNASSIGN in 189 msec 2025-01-08T02:20:44,585 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=26, resume processing ppid=24 2025-01-08T02:20:44,585 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=26, ppid=24, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithTargetName, region=77eca3710555a2734a4533147eeb6557, UNASSIGN in 191 msec 2025-01-08T02:20:44,589 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=24, resume processing ppid=23 2025-01-08T02:20:44,589 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=24, ppid=23, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithTargetName in 200 msec 2025-01-08T02:20:44,590 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302844590"}]},"ts":"1736302844590"} 2025-01-08T02:20:44,593 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithTargetName, state=DISABLED in hbase:meta 2025-01-08T02:20:44,595 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithTargetName to state=DISABLED 2025-01-08T02:20:44,598 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=23, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithTargetName in 228 msec 2025-01-08T02:20:44,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=23 2025-01-08T02:20:44,680 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithTargetName, procId: 23 completed 2025-01-08T02:20:44,684 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportWithTargetName 2025-01-08T02:20:44,691 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=29, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:44,693 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=29, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:44,695 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=29, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:44,696 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithTargetName 2025-01-08T02:20:44,702 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithTargetName 2025-01-08T02:20:44,705 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:44,706 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T02:20:44,707 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T02:20:44,708 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T02:20:44,709 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T02:20:44,709 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data PBUF 2025-01-08T02:20:44,710 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T02:20:44,710 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T02:20:44,710 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:44,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:44,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithTargetName 2025-01-08T02:20:44,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:44,711 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T02:20:44,711 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T02:20:44,713 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:44,713 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithTargetName with data null 2025-01-08T02:20:44,713 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T02:20:44,715 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=29 2025-01-08T02:20:44,715 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/recovered.edits] 2025-01-08T02:20:44,718 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:44,722 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/recovered.edits] 2025-01-08T02:20:44,727 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/cf/96698aec059d4e468c54479121710b2a to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/cf/96698aec059d4e468c54479121710b2a 2025-01-08T02:20:44,732 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/cf/9dfd847383954939bb9be9923f8f52b4 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/cf/9dfd847383954939bb9be9923f8f52b4 2025-01-08T02:20:44,736 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed/recovered.edits/9.seqid 2025-01-08T02:20:44,737 DEBUG [HFileArchiver-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:44,741 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557/recovered.edits/9.seqid 2025-01-08T02:20:44,741 DEBUG [HFileArchiver-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithTargetName/77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:44,741 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithTargetName regions 2025-01-08T02:20:44,742 DEBUG [PEWorker-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71 2025-01-08T02:20:44,744 DEBUG [PEWorker-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf] 2025-01-08T02:20:44,750 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6f537b1374b458a9364196e607979dd_77eca3710555a2734a4533147eeb6557 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/c4ca4238a0b923820dcc509a6f75849b20250108f6f537b1374b458a9364196e607979dd_77eca3710555a2734a4533147eeb6557 2025-01-08T02:20:44,751 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e20250108498e4d60a68145b18d3384bad66c27e0_260aa1a42d10faf69822c2b808eb66ed to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71/cf/d41d8cd98f00b204e9800998ecf8427e20250108498e4d60a68145b18d3384bad66c27e0_260aa1a42d10faf69822c2b808eb66ed 2025-01-08T02:20:44,752 DEBUG [PEWorker-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithTargetName/90c8eeaaf01a24f585da11044309be71 2025-01-08T02:20:44,767 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=29, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:44,774 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36743 {}] util.ReflectedFunctionCache(97): Populated cache for org.apache.hadoop.hbase.filter.KeyOnlyFilter in 0ms 2025-01-08T02:20:44,778 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithTargetName from hbase:meta 2025-01-08T02:20:44,782 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithTargetName' descriptor. 2025-01-08T02:20:44,787 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=29, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:44,787 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithTargetName' from region states. 2025-01-08T02:20:44,787 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302844787"}]},"ts":"9223372036854775807"} 2025-01-08T02:20:44,787 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302844787"}]},"ts":"9223372036854775807"} 2025-01-08T02:20:44,792 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:20:44,792 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 260aa1a42d10faf69822c2b808eb66ed, NAME => 'testtb-testExportWithTargetName,,1736302819259.260aa1a42d10faf69822c2b808eb66ed.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 77eca3710555a2734a4533147eeb6557, NAME => 'testtb-testExportWithTargetName,1,1736302819259.77eca3710555a2734a4533147eeb6557.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:20:44,792 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithTargetName' as deleted. 2025-01-08T02:20:44,792 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithTargetName","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302844792"}]},"ts":"9223372036854775807"} 2025-01-08T02:20:44,796 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithTargetName state from META 2025-01-08T02:20:44,805 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=29, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithTargetName 2025-01-08T02:20:44,813 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=29, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithTargetName in 120 msec 2025-01-08T02:20:44,816 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=29 2025-01-08T02:20:44,817 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithTargetName, procId: 29 completed 2025-01-08T02:20:44,835 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportWithTargetName" 2025-01-08T02:20:44,837 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithTargetName 2025-01-08T02:20:44,840 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportWithTargetName" 2025-01-08T02:20:44,842 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithTargetName 2025-01-08T02:20:44,890 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithTargetName Thread=768 (was 720) Potentially hanging thread: RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ContainersLauncher #0 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.read1(BufferedReader.java:213) java.base@17.0.11/java.io.BufferedReader.read(BufferedReader.java:287) app//org.apache.hadoop.util.Shell$ShellCommandExecutor.parseExecResult(Shell.java:1295) app//org.apache.hadoop.util.Shell.runCommand(Shell.java:1054) app//org.apache.hadoop.util.Shell.run(Shell.java:959) app//org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:1282) app//org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:349) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.launchContainer(ContainerLaunch.java:600) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:388) app//org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:105) java.base@17.0.11/java.util.concurrent.FutureTask.run(FutureTask.java:264) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:50578 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-1332 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: ContainersLauncher #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:35939 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:51256 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:38657 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1426171573_1 at /127.0.0.1:50544 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: htable-pool-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:35939 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: htable-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-3 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:38657 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1426171573_1 at /127.0.0.1:53076 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45347 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS-EventLoopGroup-3-3 app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 24361) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:53086 [Waiting for operation #7] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DeletionService #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Container metrics unregistration java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Potentially hanging thread: DeletionService #3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: zk-permission-watcher-pool-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-2 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Async disk worker #0 for volume /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=809 (was 781) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=815 (was 883), ProcessCount=19 (was 18) - ProcessCount LEAK? -, AvailableMemoryMB=1516 (was 1676) 2025-01-08T02:20:44,891 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=768 is superior to 500 2025-01-08T02:20:44,910 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithResetTtl Thread=768, OpenFileDescriptor=809, MaxFileDescriptor=1048576, SystemLoadAverage=815, ProcessCount=19, AvailableMemoryMB=1516 2025-01-08T02:20:44,910 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=768 is superior to 500 2025-01-08T02:20:44,913 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:20:44,914 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=30, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:20:44,915 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:20:44,916 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithResetTtl" procId is: 30 2025-01-08T02:20:44,917 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:20:44,917 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T02:20:44,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741904_1080 (size=440) 2025-01-08T02:20:44,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741904_1080 (size=440) 2025-01-08T02:20:44,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741904_1080 (size=440) 2025-01-08T02:20:44,930 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 308a78ff21bb936d4a1552dc57b4435c, NAME => 'testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:44,930 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 922e987cbaa461a827ca105c1388a10b, NAME => 'testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:44,942 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741906_1082 (size=65) 2025-01-08T02:20:44,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741906_1082 (size=65) 2025-01-08T02:20:44,945 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:44,945 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1681): Closing 922e987cbaa461a827ca105c1388a10b, disabling compactions & flushes 2025-01-08T02:20:44,945 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:44,945 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:44,945 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. after waiting 0 ms 2025-01-08T02:20:44,945 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:44,945 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:44,945 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1635): Region close journal for 922e987cbaa461a827ca105c1388a10b: 2025-01-08T02:20:44,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741906_1082 (size=65) 2025-01-08T02:20:44,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741905_1081 (size=65) 2025-01-08T02:20:44,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741905_1081 (size=65) 2025-01-08T02:20:44,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741905_1081 (size=65) 2025-01-08T02:20:45,020 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T02:20:45,221 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T02:20:45,346 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:45,346 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1681): Closing 308a78ff21bb936d4a1552dc57b4435c, disabling compactions & flushes 2025-01-08T02:20:45,346 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:45,346 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:45,346 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. after waiting 0 ms 2025-01-08T02:20:45,346 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:45,346 INFO [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:45,346 DEBUG [RegionOpenAndInit-testtb-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1635): Region close journal for 308a78ff21bb936d4a1552dc57b4435c: 2025-01-08T02:20:45,348 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:20:45,349 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736302845349"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302845349"}]},"ts":"1736302845349"} 2025-01-08T02:20:45,349 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736302845349"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302845349"}]},"ts":"1736302845349"} 2025-01-08T02:20:45,353 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:20:45,356 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:20:45,356 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302845356"}]},"ts":"1736302845356"} 2025-01-08T02:20:45,358 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=ENABLING in hbase:meta 2025-01-08T02:20:45,362 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:20:45,364 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:20:45,364 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:20:45,364 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:20:45,364 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:20:45,364 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:20:45,364 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:20:45,365 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:20:45,365 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=308a78ff21bb936d4a1552dc57b4435c, ASSIGN}, {pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=922e987cbaa461a827ca105c1388a10b, ASSIGN}] 2025-01-08T02:20:45,367 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=308a78ff21bb936d4a1552dc57b4435c, ASSIGN 2025-01-08T02:20:45,368 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=922e987cbaa461a827ca105c1388a10b, ASSIGN 2025-01-08T02:20:45,373 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=31, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=308a78ff21bb936d4a1552dc57b4435c, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:20:45,373 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=32, ppid=30, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=922e987cbaa461a827ca105c1388a10b, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:20:45,523 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T02:20:45,524 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:20:45,524 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=31 updating hbase:meta row=308a78ff21bb936d4a1552dc57b4435c, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:45,524 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=32 updating hbase:meta row=922e987cbaa461a827ca105c1388a10b, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:45,526 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=33, ppid=31, state=RUNNABLE; OpenRegionProcedure 308a78ff21bb936d4a1552dc57b4435c, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:20:45,528 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=34, ppid=32, state=RUNNABLE; OpenRegionProcedure 922e987cbaa461a827ca105c1388a10b, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:20:45,679 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:45,680 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:45,683 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] handler.AssignRegionHandler(135): Open testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:45,683 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7285): Opening region: {ENCODED => 308a78ff21bb936d4a1552dc57b4435c, NAME => 'testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:20:45,683 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] handler.AssignRegionHandler(135): Open testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:45,683 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. service=AccessControlService 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7285): Opening region: {ENCODED => 922e987cbaa461a827ca105c1388a10b, NAME => 'testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:20:45,684 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithResetTtl 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. service=AccessControlService 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7327): checking encryption for 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(7330): checking classloading for 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:45,684 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithResetTtl 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:45,684 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(894): Instantiated testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:45,685 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7327): checking encryption for 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:45,685 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(7330): checking classloading for 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:45,686 INFO [StoreOpener-308a78ff21bb936d4a1552dc57b4435c-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:45,686 INFO [StoreOpener-922e987cbaa461a827ca105c1388a10b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:45,688 INFO [StoreOpener-922e987cbaa461a827ca105c1388a10b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 922e987cbaa461a827ca105c1388a10b columnFamilyName cf 2025-01-08T02:20:45,688 INFO [StoreOpener-308a78ff21bb936d4a1552dc57b4435c-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 308a78ff21bb936d4a1552dc57b4435c columnFamilyName cf 2025-01-08T02:20:45,689 DEBUG [StoreOpener-922e987cbaa461a827ca105c1388a10b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:45,689 DEBUG [StoreOpener-308a78ff21bb936d4a1552dc57b4435c-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:45,689 INFO [StoreOpener-922e987cbaa461a827ca105c1388a10b-1 {}] regionserver.HStore(327): Store=922e987cbaa461a827ca105c1388a10b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:45,689 INFO [StoreOpener-308a78ff21bb936d4a1552dc57b4435c-1 {}] regionserver.HStore(327): Store=308a78ff21bb936d4a1552dc57b4435c/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:45,690 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:45,691 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:45,691 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:45,691 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:45,693 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1085): writing seq id for 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:45,693 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1085): writing seq id for 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:45,696 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:45,696 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:45,696 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1102): Opened 308a78ff21bb936d4a1552dc57b4435c; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63103641, jitterRate=-0.05968247354030609}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:45,696 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1102): Opened 922e987cbaa461a827ca105c1388a10b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73959444, jitterRate=0.10208159685134888}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:45,698 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegion(1001): Region open journal for 308a78ff21bb936d4a1552dc57b4435c: 2025-01-08T02:20:45,698 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegion(1001): Region open journal for 922e987cbaa461a827ca105c1388a10b: 2025-01-08T02:20:45,699 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b., pid=34, masterSystemTime=1736302845680 2025-01-08T02:20:45,699 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c., pid=33, masterSystemTime=1736302845679 2025-01-08T02:20:45,700 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:45,701 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=33}] handler.AssignRegionHandler(164): Opened testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:45,701 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=31 updating hbase:meta row=308a78ff21bb936d4a1552dc57b4435c, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:45,701 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:45,701 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=34}] handler.AssignRegionHandler(164): Opened testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:45,702 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=32 updating hbase:meta row=922e987cbaa461a827ca105c1388a10b, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:45,705 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=33, resume processing ppid=31 2025-01-08T02:20:45,705 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=33, ppid=31, state=SUCCESS; OpenRegionProcedure 308a78ff21bb936d4a1552dc57b4435c, server=c4e5d52b9759,36743,1736302808837 in 177 msec 2025-01-08T02:20:45,706 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=34, resume processing ppid=32 2025-01-08T02:20:45,706 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=34, ppid=32, state=SUCCESS; OpenRegionProcedure 922e987cbaa461a827ca105c1388a10b, server=c4e5d52b9759,37031,1736302808764 in 176 msec 2025-01-08T02:20:45,707 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=31, ppid=30, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=308a78ff21bb936d4a1552dc57b4435c, ASSIGN in 340 msec 2025-01-08T02:20:45,709 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=32, resume processing ppid=30 2025-01-08T02:20:45,709 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=32, ppid=30, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=922e987cbaa461a827ca105c1388a10b, ASSIGN in 341 msec 2025-01-08T02:20:45,709 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:20:45,710 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302845710"}]},"ts":"1736302845710"} 2025-01-08T02:20:45,711 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=ENABLED in hbase:meta 2025-01-08T02:20:45,714 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=30, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithResetTtl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:20:45,714 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithResetTtl jenkins: RWXCA 2025-01-08T02:20:45,716 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T02:20:45,718 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:45,718 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:45,718 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:45,718 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:45,721 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=30, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithResetTtl in 807 msec 2025-01-08T02:20:45,724 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:45,724 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:45,725 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:45,725 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:46,024 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=30 2025-01-08T02:20:46,024 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithResetTtl, procId: 30 completed 2025-01-08T02:20:46,028 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithResetTtl 2025-01-08T02:20:46,028 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:46,028 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:20:46,046 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T02:20:46,046 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302846046 (current time:1736302846046). 2025-01-08T02:20:46,046 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:20:46,046 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T02:20:46,046 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:20:46,047 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0832a22c to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7ee85a3 2025-01-08T02:20:46,051 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@93c5086, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:46,053 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:46,054 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41966, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:46,055 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0832a22c to 127.0.0.1:61593 2025-01-08T02:20:46,055 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:46,057 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0d20e2ca to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7e90ad50 2025-01-08T02:20:46,060 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@15436a64, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:46,062 DEBUG [hconnection-0x75b4c6c2-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:46,063 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41980, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:46,065 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:46,065 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46752, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:46,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0d20e2ca to 127.0.0.1:61593 2025-01-08T02:20:46,067 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:46,067 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T02:20:46,068 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:20:46,069 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=35, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T02:20:46,069 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 35 2025-01-08T02:20:46,070 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:20:46,070 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T02:20:46,071 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:20:46,074 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:20:46,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741907_1083 (size=161) 2025-01-08T02:20:46,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741907_1083 (size=161) 2025-01-08T02:20:46,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741907_1083 (size=161) 2025-01-08T02:20:46,085 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:20:46,085 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c}, {pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b}] 2025-01-08T02:20:46,086 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:46,086 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:46,172 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T02:20:46,238 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:46,238 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:46,239 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=36 2025-01-08T02:20:46,239 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=37 2025-01-08T02:20:46,240 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:46,240 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:46,241 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.HRegion(2538): Flush status journal for 308a78ff21bb936d4a1552dc57b4435c: 2025-01-08T02:20:46,241 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.HRegion(2538): Flush status journal for 922e987cbaa461a827ca105c1388a10b: 2025-01-08T02:20:46,241 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. for emptySnaptb0-testExportWithResetTtl completed. 2025-01-08T02:20:46,241 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.' region-info for snapshot=emptySnaptb0-testExportWithResetTtl 2025-01-08T02:20:46,241 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:46,241 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:20:46,242 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. for emptySnaptb0-testExportWithResetTtl completed. 2025-01-08T02:20:46,242 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.' region-info for snapshot=emptySnaptb0-testExportWithResetTtl 2025-01-08T02:20:46,242 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:46,242 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:20:46,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741909_1085 (size=68) 2025-01-08T02:20:46,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741909_1085 (size=68) 2025-01-08T02:20:46,271 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741909_1085 (size=68) 2025-01-08T02:20:46,272 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:46,272 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=36}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=36 2025-01-08T02:20:46,272 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=36 2025-01-08T02:20:46,273 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithResetTtl on region 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:46,273 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=36, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:46,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741908_1084 (size=68) 2025-01-08T02:20:46,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741908_1084 (size=68) 2025-01-08T02:20:46,275 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741908_1084 (size=68) 2025-01-08T02:20:46,276 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:46,276 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=37}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=37 2025-01-08T02:20:46,277 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=37 2025-01-08T02:20:46,277 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithResetTtl on region 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:46,277 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=37, ppid=35, state=RUNNABLE; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:46,279 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=36, ppid=35, state=SUCCESS; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c in 189 msec 2025-01-08T02:20:46,282 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=37, resume processing ppid=35 2025-01-08T02:20:46,283 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=37, ppid=35, state=SUCCESS; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b in 194 msec 2025-01-08T02:20:46,283 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:20:46,284 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:20:46,285 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:20:46,285 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:20:46,285 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:46,286 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:20:46,297 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741910_1086 (size=60) 2025-01-08T02:20:46,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741910_1086 (size=60) 2025-01-08T02:20:46,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741910_1086 (size=60) 2025-01-08T02:20:46,300 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:20:46,300 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithResetTtl 2025-01-08T02:20:46,301 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithResetTtl 2025-01-08T02:20:46,339 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741911_1087 (size=641) 2025-01-08T02:20:46,339 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741911_1087 (size=641) 2025-01-08T02:20:46,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741911_1087 (size=641) 2025-01-08T02:20:46,354 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:20:46,373 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T02:20:46,374 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:20:46,375 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithResetTtl to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testExportWithResetTtl 2025-01-08T02:20:46,377 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=35, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:20:46,377 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 35 2025-01-08T02:20:46,379 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=35, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=35, snapshot={ ss=emptySnaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } in 309 msec 2025-01-08T02:20:46,675 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=35 2025-01-08T02:20:46,675 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithResetTtl, procId: 35 completed 2025-01-08T02:20:46,683 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:20:46,688 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:20:46,695 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithResetTtl 2025-01-08T02:20:46,695 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:46,695 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:20:46,726 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T02:20:46,726 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302846726 (current time:1736302846726). 2025-01-08T02:20:46,726 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:20:46,726 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T02:20:46,726 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:20:46,728 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0a17de2f to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@59c69b79 2025-01-08T02:20:46,741 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@64135471, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:46,746 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:46,748 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41982, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:46,749 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0a17de2f to 127.0.0.1:61593 2025-01-08T02:20:46,749 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:46,751 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0b1a8b49 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@10fcae94 2025-01-08T02:20:46,765 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2015014b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:46,773 DEBUG [hconnection-0x556893d3-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:46,774 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41998, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:46,776 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:46,777 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46754, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:46,779 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0b1a8b49 to 127.0.0.1:61593 2025-01-08T02:20:46,779 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:46,780 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T02:20:46,784 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:20:46,801 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=38, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } 2025-01-08T02:20:46,802 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 38 2025-01-08T02:20:46,806 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T02:20:46,823 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:20:46,829 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:20:46,841 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:20:46,852 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741912_1088 (size=156) 2025-01-08T02:20:46,852 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741912_1088 (size=156) 2025-01-08T02:20:46,856 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741912_1088 (size=156) 2025-01-08T02:20:46,859 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:20:46,859 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c}, {pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b}] 2025-01-08T02:20:46,861 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:46,861 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:46,912 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithResetTtl' 2025-01-08T02:20:46,922 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T02:20:47,013 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:20:47,014 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:47,015 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=39 2025-01-08T02:20:47,015 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:47,015 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(2837): Flushing 308a78ff21bb936d4a1552dc57b4435c 1/1 column families, dataSize=132 B heapSize=544 B 2025-01-08T02:20:47,016 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=40 2025-01-08T02:20:47,016 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:47,016 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(2837): Flushing 922e987cbaa461a827ca105c1388a10b 1/1 column families, dataSize=3.13 KB heapSize=7 KB 2025-01-08T02:20:47,055 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108683b53efe767410a96bccb9d35cd1e4f_308a78ff21bb936d4a1552dc57b4435c is 71, key is 046794047acb9792726e60049082a6f4/cf:q/1736302846683/Put/seqid=0 2025-01-08T02:20:47,062 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010884eee7384bab40bcb6da79600695569b_922e987cbaa461a827ca105c1388a10b is 71, key is 11f8943edf32e5df6d43ba2d834ff226/cf:q/1736302846688/Put/seqid=0 2025-01-08T02:20:47,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741913_1089 (size=5032) 2025-01-08T02:20:47,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741913_1089 (size=5032) 2025-01-08T02:20:47,094 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741913_1089 (size=5032) 2025-01-08T02:20:47,095 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:47,106 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741914_1090 (size=8241) 2025-01-08T02:20:47,106 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741914_1090 (size=8241) 2025-01-08T02:20:47,110 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741914_1090 (size=8241) 2025-01-08T02:20:47,111 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:47,114 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108683b53efe767410a96bccb9d35cd1e4f_308a78ff21bb936d4a1552dc57b4435c to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108683b53efe767410a96bccb9d35cd1e4f_308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:47,116 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/.tmp/cf/0e0b7b1f2bb64e05a22e50c20ad43616, store: [table=testtb-testExportWithResetTtl family=cf region=308a78ff21bb936d4a1552dc57b4435c] 2025-01-08T02:20:47,122 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010884eee7384bab40bcb6da79600695569b_922e987cbaa461a827ca105c1388a10b to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010884eee7384bab40bcb6da79600695569b_922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:47,124 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/.tmp/cf/eeca5b5223ab4053b46a1ed3239aa3a2, store: [table=testtb-testExportWithResetTtl family=cf region=922e987cbaa461a827ca105c1388a10b] 2025-01-08T02:20:47,124 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T02:20:47,124 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/.tmp/cf/0e0b7b1f2bb64e05a22e50c20ad43616 is 206, key is 069ac58d05df0f5e9d9805e41e817f503/cf:q/1736302846683/Put/seqid=0 2025-01-08T02:20:47,125 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/.tmp/cf/eeca5b5223ab4053b46a1ed3239aa3a2 is 206, key is 15dd17c5ca9dba237d870259aafe2b69a/cf:q/1736302846688/Put/seqid=0 2025-01-08T02:20:47,134 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741915_1091 (size=5700) 2025-01-08T02:20:47,134 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741915_1091 (size=5700) 2025-01-08T02:20:47,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741915_1091 (size=5700) 2025-01-08T02:20:47,135 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=132, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/.tmp/cf/0e0b7b1f2bb64e05a22e50c20ad43616 2025-01-08T02:20:47,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741916_1092 (size=15055) 2025-01-08T02:20:47,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741916_1092 (size=15055) 2025-01-08T02:20:47,138 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741916_1092 (size=15055) 2025-01-08T02:20:47,139 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/.tmp/cf/eeca5b5223ab4053b46a1ed3239aa3a2 2025-01-08T02:20:47,144 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/.tmp/cf/0e0b7b1f2bb64e05a22e50c20ad43616 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/cf/0e0b7b1f2bb64e05a22e50c20ad43616 2025-01-08T02:20:47,147 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/.tmp/cf/eeca5b5223ab4053b46a1ed3239aa3a2 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/cf/eeca5b5223ab4053b46a1ed3239aa3a2 2025-01-08T02:20:47,153 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/cf/0e0b7b1f2bb64e05a22e50c20ad43616, entries=2, sequenceid=6, filesize=5.6 K 2025-01-08T02:20:47,154 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(3040): Finished flush of dataSize ~132 B/132, heapSize ~528 B/528, currentSize=0 B/0 for 308a78ff21bb936d4a1552dc57b4435c in 139ms, sequenceid=6, compaction requested=false 2025-01-08T02:20:47,154 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.HRegion(2538): Flush status journal for 308a78ff21bb936d4a1552dc57b4435c: 2025-01-08T02:20:47,154 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/cf/eeca5b5223ab4053b46a1ed3239aa3a2, entries=48, sequenceid=6, filesize=14.7 K 2025-01-08T02:20:47,154 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. for snaptb0-testExportWithResetTtl completed. 2025-01-08T02:20:47,154 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.' region-info for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T02:20:47,154 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:47,154 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/cf/0e0b7b1f2bb64e05a22e50c20ad43616] hfiles 2025-01-08T02:20:47,154 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/cf/0e0b7b1f2bb64e05a22e50c20ad43616 for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T02:20:47,155 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(3040): Finished flush of dataSize ~3.13 KB/3204, heapSize ~6.98 KB/7152, currentSize=0 B/0 for 922e987cbaa461a827ca105c1388a10b in 139ms, sequenceid=6, compaction requested=false 2025-01-08T02:20:47,156 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.HRegion(2538): Flush status journal for 922e987cbaa461a827ca105c1388a10b: 2025-01-08T02:20:47,156 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. for snaptb0-testExportWithResetTtl completed. 2025-01-08T02:20:47,156 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.' region-info for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T02:20:47,156 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:47,156 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/cf/eeca5b5223ab4053b46a1ed3239aa3a2] hfiles 2025-01-08T02:20:47,156 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/cf/eeca5b5223ab4053b46a1ed3239aa3a2 for snapshot=snaptb0-testExportWithResetTtl 2025-01-08T02:20:47,170 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741917_1093 (size=107) 2025-01-08T02:20:47,177 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741918_1094 (size=107) 2025-01-08T02:20:47,177 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741918_1094 (size=107) 2025-01-08T02:20:47,178 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741918_1094 (size=107) 2025-01-08T02:20:47,178 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741917_1093 (size=107) 2025-01-08T02:20:47,178 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741917_1093 (size=107) 2025-01-08T02:20:47,180 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:20:47,181 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=39}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=39 2025-01-08T02:20:47,181 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=39 2025-01-08T02:20:47,181 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithResetTtl on region 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:47,181 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=39, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:47,185 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=39, ppid=38, state=SUCCESS; SnapshotRegionProcedure 308a78ff21bb936d4a1552dc57b4435c in 324 msec 2025-01-08T02:20:47,429 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T02:20:47,579 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:20:47,579 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=40}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=40 2025-01-08T02:20:47,580 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=40 2025-01-08T02:20:47,580 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithResetTtl on region 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:47,580 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=40, ppid=38, state=RUNNABLE; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:47,587 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=40, resume processing ppid=38 2025-01-08T02:20:47,587 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=40, ppid=38, state=SUCCESS; SnapshotRegionProcedure 922e987cbaa461a827ca105c1388a10b in 723 msec 2025-01-08T02:20:47,588 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:20:47,589 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:20:47,594 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:20:47,594 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:20:47,594 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:47,596 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010884eee7384bab40bcb6da79600695569b_922e987cbaa461a827ca105c1388a10b, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108683b53efe767410a96bccb9d35cd1e4f_308a78ff21bb936d4a1552dc57b4435c] hfiles 2025-01-08T02:20:47,597 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010884eee7384bab40bcb6da79600695569b_922e987cbaa461a827ca105c1388a10b 2025-01-08T02:20:47,597 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108683b53efe767410a96bccb9d35cd1e4f_308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:20:47,616 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741919_1095 (size=291) 2025-01-08T02:20:47,616 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741919_1095 (size=291) 2025-01-08T02:20:47,617 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741919_1095 (size=291) 2025-01-08T02:20:47,619 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:20:47,619 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithResetTtl 2025-01-08T02:20:47,620 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportWithResetTtl 2025-01-08T02:20:47,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741920_1096 (size=951) 2025-01-08T02:20:47,656 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741920_1096 (size=951) 2025-01-08T02:20:47,658 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741920_1096 (size=951) 2025-01-08T02:20:47,675 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:20:47,696 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:20:47,697 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportWithResetTtl to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithResetTtl 2025-01-08T02:20:47,699 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=38, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:20:47,699 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 }, snapshot procedure id = 38 2025-01-08T02:20:47,702 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=38, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=38, snapshot={ ss=snaptb0-testExportWithResetTtl table=testtb-testExportWithResetTtl type=FLUSH ttl=0 } in 915 msec 2025-01-08T02:20:47,931 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=38 2025-01-08T02:20:47,931 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithResetTtl, procId: 38 completed 2025-01-08T02:20:47,933 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:20:47,934 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=41, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testExportWithResetTtl 2025-01-08T02:20:47,936 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:20:47,936 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testExportWithResetTtl" procId is: 41 2025-01-08T02:20:47,937 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:20:47,937 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T02:20:47,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741921_1097 (size=433) 2025-01-08T02:20:47,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741921_1097 (size=433) 2025-01-08T02:20:47,948 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741921_1097 (size=433) 2025-01-08T02:20:47,950 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => ebf5e7819c73f9c4f5923fe9cbc7ef7f, NAME => 'testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:47,950 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 7c90cc4cbd2d273c4f5ab0dde7177718, NAME => 'testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testExportWithResetTtl', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:47,964 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741922_1098 (size=58) 2025-01-08T02:20:47,964 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741922_1098 (size=58) 2025-01-08T02:20:47,966 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:47,966 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1681): Closing ebf5e7819c73f9c4f5923fe9cbc7ef7f, disabling compactions & flushes 2025-01-08T02:20:47,971 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:47,971 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741923_1099 (size=58) 2025-01-08T02:20:47,972 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:47,972 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. after waiting 0 ms 2025-01-08T02:20:47,972 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:47,972 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1922): Closed testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:47,972 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-0 {}] regionserver.HRegion(1635): Region close journal for ebf5e7819c73f9c4f5923fe9cbc7ef7f: 2025-01-08T02:20:47,972 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741923_1099 (size=58) 2025-01-08T02:20:47,972 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741922_1098 (size=58) 2025-01-08T02:20:47,972 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:47,972 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741923_1099 (size=58) 2025-01-08T02:20:47,973 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1681): Closing 7c90cc4cbd2d273c4f5ab0dde7177718, disabling compactions & flushes 2025-01-08T02:20:47,973 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:47,973 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:47,973 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. after waiting 0 ms 2025-01-08T02:20:47,973 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:47,973 INFO [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1922): Closed testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:47,973 DEBUG [RegionOpenAndInit-testExportWithResetTtl-pool-1 {}] regionserver.HRegion(1635): Region close journal for 7c90cc4cbd2d273c4f5ab0dde7177718: 2025-01-08T02:20:47,976 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:20:47,976 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.","families":{"info":[{"qualifier":"regioninfo","vlen":57,"tag":[],"timestamp":"1736302847976"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302847976"}]},"ts":"1736302847976"} 2025-01-08T02:20:47,976 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.","families":{"info":[{"qualifier":"regioninfo","vlen":57,"tag":[],"timestamp":"1736302847976"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302847976"}]},"ts":"1736302847976"} 2025-01-08T02:20:47,984 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:20:47,988 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:20:47,989 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302847988"}]},"ts":"1736302847988"} 2025-01-08T02:20:47,991 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=ENABLING in hbase:meta 2025-01-08T02:20:47,994 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:20:47,997 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:20:47,997 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:20:47,997 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:20:47,997 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:20:47,997 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:20:47,997 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:20:47,997 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:20:47,997 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=ebf5e7819c73f9c4f5923fe9cbc7ef7f, ASSIGN}, {pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=7c90cc4cbd2d273c4f5ab0dde7177718, ASSIGN}] 2025-01-08T02:20:47,998 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=7c90cc4cbd2d273c4f5ab0dde7177718, ASSIGN 2025-01-08T02:20:47,999 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportWithResetTtl, region=ebf5e7819c73f9c4f5923fe9cbc7ef7f, ASSIGN 2025-01-08T02:20:48,000 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=43, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportWithResetTtl, region=7c90cc4cbd2d273c4f5ab0dde7177718, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:20:48,000 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=42, ppid=41, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportWithResetTtl, region=ebf5e7819c73f9c4f5923fe9cbc7ef7f, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:20:48,038 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T02:20:48,150 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:20:48,151 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=43 updating hbase:meta row=7c90cc4cbd2d273c4f5ab0dde7177718, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:48,151 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=42 updating hbase:meta row=ebf5e7819c73f9c4f5923fe9cbc7ef7f, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:48,155 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=44, ppid=43, state=RUNNABLE; OpenRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:20:48,156 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=45, ppid=42, state=RUNNABLE; OpenRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:20:48,241 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T02:20:48,308 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:48,308 DEBUG [RSProcedureDispatcher-pool-1 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T02:20:48,309 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:48,312 INFO [RS-EventLoopGroup-5-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:53524, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T02:20:48,314 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] handler.AssignRegionHandler(135): Open testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:48,314 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7285): Opening region: {ENCODED => ebf5e7819c73f9c4f5923fe9cbc7ef7f, NAME => 'testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:20:48,316 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. service=AccessControlService 2025-01-08T02:20:48,316 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:48,316 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportWithResetTtl ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,316 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:48,317 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7327): checking encryption for ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,317 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(7330): checking classloading for ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,320 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] handler.AssignRegionHandler(135): Open testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:48,320 INFO [StoreOpener-ebf5e7819c73f9c4f5923fe9cbc7ef7f-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,320 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7285): Opening region: {ENCODED => 7c90cc4cbd2d273c4f5ab0dde7177718, NAME => 'testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:20:48,321 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. service=AccessControlService 2025-01-08T02:20:48,321 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:20:48,321 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportWithResetTtl 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,321 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(894): Instantiated testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:20:48,322 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7327): checking encryption for 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,322 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(7330): checking classloading for 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,327 INFO [StoreOpener-ebf5e7819c73f9c4f5923fe9cbc7ef7f-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region ebf5e7819c73f9c4f5923fe9cbc7ef7f columnFamilyName cf 2025-01-08T02:20:48,329 DEBUG [StoreOpener-ebf5e7819c73f9c4f5923fe9cbc7ef7f-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:48,330 INFO [StoreOpener-ebf5e7819c73f9c4f5923fe9cbc7ef7f-1 {}] regionserver.HStore(327): Store=ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:48,332 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,332 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,337 INFO [StoreOpener-7c90cc4cbd2d273c4f5ab0dde7177718-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,338 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1085): writing seq id for ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,344 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:48,345 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1102): Opened ebf5e7819c73f9c4f5923fe9cbc7ef7f; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65201406, jitterRate=-0.028423339128494263}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:48,346 INFO [StoreOpener-7c90cc4cbd2d273c4f5ab0dde7177718-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 7c90cc4cbd2d273c4f5ab0dde7177718 columnFamilyName cf 2025-01-08T02:20:48,347 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegion(1001): Region open journal for ebf5e7819c73f9c4f5923fe9cbc7ef7f: 2025-01-08T02:20:48,347 DEBUG [StoreOpener-7c90cc4cbd2d273c4f5ab0dde7177718-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:48,347 INFO [StoreOpener-7c90cc4cbd2d273c4f5ab0dde7177718-1 {}] regionserver.HStore(327): Store=7c90cc4cbd2d273c4f5ab0dde7177718/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:20:48,348 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f., pid=45, masterSystemTime=1736302848309 2025-01-08T02:20:48,350 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,352 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,357 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=42 updating hbase:meta row=ebf5e7819c73f9c4f5923fe9cbc7ef7f, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:48,357 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:48,357 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=45}] handler.AssignRegionHandler(164): Opened testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:48,359 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1085): writing seq id for 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,363 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=45, resume processing ppid=42 2025-01-08T02:20:48,363 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=45, ppid=42, state=SUCCESS; OpenRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f, server=c4e5d52b9759,36743,1736302808837 in 203 msec 2025-01-08T02:20:48,364 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=42, ppid=41, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=ebf5e7819c73f9c4f5923fe9cbc7ef7f, ASSIGN in 366 msec 2025-01-08T02:20:48,372 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:20:48,374 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1102): Opened 7c90cc4cbd2d273c4f5ab0dde7177718; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70473658, jitterRate=0.05013933777809143}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:20:48,375 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegion(1001): Region open journal for 7c90cc4cbd2d273c4f5ab0dde7177718: 2025-01-08T02:20:48,376 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718., pid=44, masterSystemTime=1736302848308 2025-01-08T02:20:48,378 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:48,378 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=44}] handler.AssignRegionHandler(164): Opened testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:48,379 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=43 updating hbase:meta row=7c90cc4cbd2d273c4f5ab0dde7177718, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:48,385 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=44, resume processing ppid=43 2025-01-08T02:20:48,385 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=44, ppid=43, state=SUCCESS; OpenRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718, server=c4e5d52b9759,39911,1736302808913 in 227 msec 2025-01-08T02:20:48,389 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=43, resume processing ppid=41 2025-01-08T02:20:48,389 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=43, ppid=41, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=7c90cc4cbd2d273c4f5ab0dde7177718, ASSIGN in 388 msec 2025-01-08T02:20:48,391 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:20:48,391 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302848391"}]},"ts":"1736302848391"} 2025-01-08T02:20:48,393 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=ENABLED in hbase:meta 2025-01-08T02:20:48,397 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=41, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testExportWithResetTtl execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:20:48,398 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testExportWithResetTtl jenkins: RWXCA 2025-01-08T02:20:48,401 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T02:20:48,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:48,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:48,403 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:48,405 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:20:48,406 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,406 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,407 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,407 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,407 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,407 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,409 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=41, state=SUCCESS; CreateTableProcedure table=testExportWithResetTtl in 473 msec 2025-01-08T02:20:48,410 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,410 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF\x0A<\x0A\x07jenkins\x121\x08\x03"-\x0A!\x0A\x07default\x12\x16testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:20:48,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl 2025-01-08T02:20:48,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl Metrics about Tables on a single HBase RegionServer 2025-01-08T02:20:48,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl 2025-01-08T02:20:48,432 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl Metrics about Tables on a single HBase RegionServer 2025-01-08T02:20:48,433 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithTargetName 2025-01-08T02:20:48,548 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=41 2025-01-08T02:20:48,549 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testExportWithResetTtl, procId: 41 completed 2025-01-08T02:20:48,554 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportWithResetTtl 2025-01-08T02:20:48,554 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:48,555 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:20:48,574 DEBUG [hconnection-0x510316c8-shared-pool-11 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:48,576 INFO [RS-EventLoopGroup-5-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:53526, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:48,583 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.HRegion(8254): writing data to region testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:20:48,585 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:20:48,589 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportWithResetTtl 2025-01-08T02:20:48,589 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:48,590 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:20:48,605 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } 2025-01-08T02:20:48,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302848605 (current time:1736302848605). 2025-01-08T02:20:48,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb-testExportWithResetTtl VERSION not specified, setting to 2 2025-01-08T02:20:48,605 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:20:48,607 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4f1219e3 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@a109e65 2025-01-08T02:20:48,611 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1cf80b5e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:48,612 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:48,614 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42010, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:48,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4f1219e3 to 127.0.0.1:61593 2025-01-08T02:20:48,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:48,616 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x64523873 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6a3f612a 2025-01-08T02:20:48,621 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@36c125c8, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:20:48,622 DEBUG [hconnection-0xa32d7c0-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:48,623 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:42022, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:48,626 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:20:48,627 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46760, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:20:48,628 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x64523873 to 127.0.0.1:61593 2025-01-08T02:20:48,628 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:20:48,628 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testExportWithResetTtl], kv [jenkins: RWXCA] 2025-01-08T02:20:48,629 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:20:48,630 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=46, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } 2025-01-08T02:20:48,630 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 }, snapshot procedure id = 46 2025-01-08T02:20:48,631 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T02:20:48,631 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:20:48,633 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:20:48,636 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:20:48,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741924_1100 (size=143) 2025-01-08T02:20:48,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741924_1100 (size=143) 2025-01-08T02:20:48,655 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741924_1100 (size=143) 2025-01-08T02:20:48,660 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:20:48,661 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f}, {pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718}] 2025-01-08T02:20:48,662 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,662 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,733 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T02:20:48,813 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:20:48,813 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:20:48,814 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=47 2025-01-08T02:20:48,814 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=48 2025-01-08T02:20:48,814 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:48,815 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(2837): Flushing ebf5e7819c73f9c4f5923fe9cbc7ef7f 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T02:20:48,815 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:48,815 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(2837): Flushing 7c90cc4cbd2d273c4f5ab0dde7177718 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T02:20:48,843 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010878e9ef2ff26e417a9d84db55b95e89ec_ebf5e7819c73f9c4f5923fe9cbc7ef7f is 71, key is 0c931e78f7952a1fdbec3612ac21e940/cf:q/1736302848583/Put/seqid=0 2025-01-08T02:20:48,851 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501086794d754f7d944c78f9e8e3a597f8e1e_7c90cc4cbd2d273c4f5ab0dde7177718 is 71, key is 1119ae8f14748428434f817bbcb53df1/cf:q/1736302848585/Put/seqid=0 2025-01-08T02:20:48,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741925_1101 (size=5102) 2025-01-08T02:20:48,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741925_1101 (size=5102) 2025-01-08T02:20:48,864 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741925_1101 (size=5102) 2025-01-08T02:20:48,865 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:48,874 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010878e9ef2ff26e417a9d84db55b95e89ec_ebf5e7819c73f9c4f5923fe9cbc7ef7f to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e2025010878e9ef2ff26e417a9d84db55b95e89ec_ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,877 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/.tmp/cf/94a528e7e3454ae4849e138700b54e86, store: [table=testExportWithResetTtl family=cf region=ebf5e7819c73f9c4f5923fe9cbc7ef7f] 2025-01-08T02:20:48,878 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/.tmp/cf/94a528e7e3454ae4849e138700b54e86 is 199, key is 05c9dca9f822838144639491bb1d35ed3/cf:q/1736302848583/Put/seqid=0 2025-01-08T02:20:48,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741926_1102 (size=8171) 2025-01-08T02:20:48,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741926_1102 (size=8171) 2025-01-08T02:20:48,885 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741926_1102 (size=8171) 2025-01-08T02:20:48,885 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:48,906 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741927_1103 (size=5878) 2025-01-08T02:20:48,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741927_1103 (size=5878) 2025-01-08T02:20:48,908 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741927_1103 (size=5878) 2025-01-08T02:20:48,909 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/.tmp/cf/94a528e7e3454ae4849e138700b54e86 2025-01-08T02:20:48,912 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0001_000001 (auth:SIMPLE) from 127.0.0.1:51104 2025-01-08T02:20:48,912 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501086794d754f7d944c78f9e8e3a597f8e1e_7c90cc4cbd2d273c4f5ab0dde7177718 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b202501086794d754f7d944c78f9e8e3a597f8e1e_7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,919 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_3/usercache/jenkins/appcache/application_1736302816589_0001/container_1736302816589_0001_01_000001/launch_container.sh] 2025-01-08T02:20:48,919 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_3/usercache/jenkins/appcache/application_1736302816589_0001/container_1736302816589_0001_01_000001/container_tokens] 2025-01-08T02:20:48,919 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_3/usercache/jenkins/appcache/application_1736302816589_0001/container_1736302816589_0001_01_000001/sysfs] 2025-01-08T02:20:48,919 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/.tmp/cf/3457da62efa340e1814230be8e434b44, store: [table=testExportWithResetTtl family=cf region=7c90cc4cbd2d273c4f5ab0dde7177718] 2025-01-08T02:20:48,920 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/.tmp/cf/3457da62efa340e1814230be8e434b44 is 199, key is 1e67ab7449431b368fd61586b7f94ef90/cf:q/1736302848585/Put/seqid=0 2025-01-08T02:20:48,922 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/.tmp/cf/94a528e7e3454ae4849e138700b54e86 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf/94a528e7e3454ae4849e138700b54e86 2025-01-08T02:20:48,930 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf/94a528e7e3454ae4849e138700b54e86, entries=3, sequenceid=5, filesize=5.7 K 2025-01-08T02:20:48,931 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for ebf5e7819c73f9c4f5923fe9cbc7ef7f in 116ms, sequenceid=5, compaction requested=false 2025-01-08T02:20:48,931 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testExportWithResetTtl' 2025-01-08T02:20:48,932 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.HRegion(2538): Flush status journal for ebf5e7819c73f9c4f5923fe9cbc7ef7f: 2025-01-08T02:20:48,932 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. for snaptb-testExportWithResetTtl completed. 2025-01-08T02:20:48,932 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(241): Storing 'testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.' region-info for snapshot=snaptb-testExportWithResetTtl 2025-01-08T02:20:48,932 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:48,932 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf/94a528e7e3454ae4849e138700b54e86] hfiles 2025-01-08T02:20:48,932 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf/94a528e7e3454ae4849e138700b54e86 for snapshot=snaptb-testExportWithResetTtl 2025-01-08T02:20:48,934 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T02:20:48,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741928_1104 (size=14517) 2025-01-08T02:20:48,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741928_1104 (size=14517) 2025-01-08T02:20:48,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741928_1104 (size=14517) 2025-01-08T02:20:48,956 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/.tmp/cf/3457da62efa340e1814230be8e434b44 2025-01-08T02:20:48,964 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/.tmp/cf/3457da62efa340e1814230be8e434b44 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/cf/3457da62efa340e1814230be8e434b44 2025-01-08T02:20:48,965 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741929_1105 (size=100) 2025-01-08T02:20:48,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741929_1105 (size=100) 2025-01-08T02:20:48,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741929_1105 (size=100) 2025-01-08T02:20:48,967 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:20:48,967 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=47}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=47 2025-01-08T02:20:48,968 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=47 2025-01-08T02:20:48,968 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb-testExportWithResetTtl on region ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,968 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=47, ppid=46, state=RUNNABLE; SnapshotRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:48,972 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/cf/3457da62efa340e1814230be8e434b44, entries=47, sequenceid=5, filesize=14.2 K 2025-01-08T02:20:48,973 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for 7c90cc4cbd2d273c4f5ab0dde7177718 in 158ms, sequenceid=5, compaction requested=false 2025-01-08T02:20:48,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.HRegion(2538): Flush status journal for 7c90cc4cbd2d273c4f5ab0dde7177718: 2025-01-08T02:20:48,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. for snaptb-testExportWithResetTtl completed. 2025-01-08T02:20:48,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(241): Storing 'testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.' region-info for snapshot=snaptb-testExportWithResetTtl 2025-01-08T02:20:48,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:20:48,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/cf/3457da62efa340e1814230be8e434b44] hfiles 2025-01-08T02:20:48,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/cf/3457da62efa340e1814230be8e434b44 for snapshot=snaptb-testExportWithResetTtl 2025-01-08T02:20:48,977 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=47, ppid=46, state=SUCCESS; SnapshotRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f in 309 msec 2025-01-08T02:20:48,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741930_1106 (size=100) 2025-01-08T02:20:48,986 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741930_1106 (size=100) 2025-01-08T02:20:48,988 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741930_1106 (size=100) 2025-01-08T02:20:48,989 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:20:48,989 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=48}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=48 2025-01-08T02:20:48,990 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=48 2025-01-08T02:20:48,990 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb-testExportWithResetTtl on region 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,990 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=48, ppid=46, state=RUNNABLE; SnapshotRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,993 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=48, resume processing ppid=46 2025-01-08T02:20:48,993 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=48, ppid=46, state=SUCCESS; SnapshotRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718 in 330 msec 2025-01-08T02:20:48,993 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:20:48,993 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:20:48,995 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:20:48,995 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:20:48,995 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:20:48,996 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b202501086794d754f7d944c78f9e8e3a597f8e1e_7c90cc4cbd2d273c4f5ab0dde7177718, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e2025010878e9ef2ff26e417a9d84db55b95e89ec_ebf5e7819c73f9c4f5923fe9cbc7ef7f] hfiles 2025-01-08T02:20:48,996 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b202501086794d754f7d944c78f9e8e3a597f8e1e_7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:20:48,996 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e2025010878e9ef2ff26e417a9d84db55b95e89ec_ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:20:49,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741931_1107 (size=284) 2025-01-08T02:20:49,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741931_1107 (size=284) 2025-01-08T02:20:49,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741931_1107 (size=284) 2025-01-08T02:20:49,007 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:20:49,007 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb-testExportWithResetTtl 2025-01-08T02:20:49,007 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T02:20:49,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741932_1108 (size=923) 2025-01-08T02:20:49,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741932_1108 (size=923) 2025-01-08T02:20:49,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741932_1108 (size=923) 2025-01-08T02:20:49,039 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:20:49,051 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:20:49,051 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T02:20:49,054 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=46, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:20:49,054 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 }, snapshot procedure id = 46 2025-01-08T02:20:49,056 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=46, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=46, snapshot={ ss=snaptb-testExportWithResetTtl table=testExportWithResetTtl type=FLUSH ttl=100000 } in 425 msec 2025-01-08T02:20:49,235 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=46 2025-01-08T02:20:49,236 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testExportWithResetTtl, procId: 46 completed 2025-01-08T02:20:49,248 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248 2025-01-08T02:20:49,248 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:49,298 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:20:49,298 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T02:20:49,301 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:20:49,307 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb-testExportWithResetTtl to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248/.hbase-snapshot/.tmp/snaptb-testExportWithResetTtl 2025-01-08T02:20:49,320 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741933_1109 (size=923) 2025-01-08T02:20:49,320 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741933_1109 (size=923) 2025-01-08T02:20:49,320 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741933_1109 (size=923) 2025-01-08T02:20:49,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741934_1110 (size=143) 2025-01-08T02:20:49,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741934_1110 (size=143) 2025-01-08T02:20:49,327 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741934_1110 (size=143) 2025-01-08T02:20:49,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741935_1111 (size=141) 2025-01-08T02:20:49,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741935_1111 (size=141) 2025-01-08T02:20:49,337 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741935_1111 (size=141) 2025-01-08T02:20:49,339 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:49,339 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:49,340 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:49,340 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,120 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:20:50,371 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-10135823373980271664.jar 2025-01-08T02:20:50,372 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,372 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,443 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-6170620627724780890.jar 2025-01-08T02:20:50,444 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,444 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,444 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,445 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,445 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,445 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:20:50,445 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:20:50,446 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:20:50,446 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:20:50,447 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:20:50,447 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:20:50,447 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:20:50,447 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:20:50,448 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:20:50,448 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:20:50,448 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:20:50,448 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:20:50,449 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:20:50,449 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:50,449 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:50,449 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:20:50,450 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:50,450 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:20:50,450 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:20:50,450 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:20:50,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741936_1112 (size=127628) 2025-01-08T02:20:50,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741936_1112 (size=127628) 2025-01-08T02:20:50,510 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741936_1112 (size=127628) 2025-01-08T02:20:50,531 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T02:20:50,532 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T02:20:50,532 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741937_1113 (size=2172137) 2025-01-08T02:20:50,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741938_1114 (size=213228) 2025-01-08T02:20:50,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741938_1114 (size=213228) 2025-01-08T02:20:50,548 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741938_1114 (size=213228) 2025-01-08T02:20:50,572 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T02:20:50,573 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T02:20:50,573 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741939_1115 (size=1877034) 2025-01-08T02:20:50,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741940_1116 (size=533455) 2025-01-08T02:20:50,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741940_1116 (size=533455) 2025-01-08T02:20:50,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741940_1116 (size=533455) 2025-01-08T02:20:50,627 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T02:20:50,627 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T02:20:50,628 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741941_1117 (size=7280644) 2025-01-08T02:20:50,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741942_1118 (size=451756) 2025-01-08T02:20:50,643 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741942_1118 (size=451756) 2025-01-08T02:20:50,643 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741942_1118 (size=451756) 2025-01-08T02:20:50,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741943_1119 (size=4188619) 2025-01-08T02:20:50,685 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741943_1119 (size=4188619) 2025-01-08T02:20:50,686 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741943_1119 (size=4188619) 2025-01-08T02:20:50,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741944_1120 (size=20406) 2025-01-08T02:20:50,695 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741944_1120 (size=20406) 2025-01-08T02:20:50,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741944_1120 (size=20406) 2025-01-08T02:20:50,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741945_1121 (size=75495) 2025-01-08T02:20:50,704 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741945_1121 (size=75495) 2025-01-08T02:20:50,705 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741945_1121 (size=75495) 2025-01-08T02:20:50,713 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741946_1122 (size=45609) 2025-01-08T02:20:50,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741946_1122 (size=45609) 2025-01-08T02:20:50,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741946_1122 (size=45609) 2025-01-08T02:20:50,724 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741947_1123 (size=110084) 2025-01-08T02:20:50,724 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741947_1123 (size=110084) 2025-01-08T02:20:50,725 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741947_1123 (size=110084) 2025-01-08T02:20:50,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741948_1124 (size=1323991) 2025-01-08T02:20:50,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741948_1124 (size=1323991) 2025-01-08T02:20:50,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741948_1124 (size=1323991) 2025-01-08T02:20:50,748 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741949_1125 (size=23076) 2025-01-08T02:20:50,749 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741949_1125 (size=23076) 2025-01-08T02:20:50,749 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741949_1125 (size=23076) 2025-01-08T02:20:50,759 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741950_1126 (size=126803) 2025-01-08T02:20:50,760 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741950_1126 (size=126803) 2025-01-08T02:20:50,760 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741950_1126 (size=126803) 2025-01-08T02:20:50,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741951_1127 (size=322274) 2025-01-08T02:20:50,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741951_1127 (size=322274) 2025-01-08T02:20:50,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741951_1127 (size=322274) 2025-01-08T02:20:50,840 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741952_1128 (size=1832290) 2025-01-08T02:20:50,840 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741952_1128 (size=1832290) 2025-01-08T02:20:50,841 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741952_1128 (size=1832290) 2025-01-08T02:20:50,925 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741953_1129 (size=6350715) 2025-01-08T02:20:50,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741953_1129 (size=6350715) 2025-01-08T02:20:50,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741953_1129 (size=6350715) 2025-01-08T02:20:50,942 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741954_1130 (size=30081) 2025-01-08T02:20:50,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741954_1130 (size=30081) 2025-01-08T02:20:50,943 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741954_1130 (size=30081) 2025-01-08T02:20:50,959 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741955_1131 (size=53616) 2025-01-08T02:20:50,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741955_1131 (size=53616) 2025-01-08T02:20:50,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741955_1131 (size=53616) 2025-01-08T02:20:51,009 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741956_1132 (size=29229) 2025-01-08T02:20:51,010 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741956_1132 (size=29229) 2025-01-08T02:20:51,010 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741956_1132 (size=29229) 2025-01-08T02:20:51,068 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741957_1133 (size=169089) 2025-01-08T02:20:51,070 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741957_1133 (size=169089) 2025-01-08T02:20:51,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741957_1133 (size=169089) 2025-01-08T02:20:51,157 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741958_1134 (size=5175431) 2025-01-08T02:20:51,158 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741958_1134 (size=5175431) 2025-01-08T02:20:51,158 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741958_1134 (size=5175431) 2025-01-08T02:20:51,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741959_1135 (size=136454) 2025-01-08T02:20:51,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741959_1135 (size=136454) 2025-01-08T02:20:51,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741959_1135 (size=136454) 2025-01-08T02:20:51,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741960_1136 (size=907469) 2025-01-08T02:20:51,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741960_1136 (size=907469) 2025-01-08T02:20:51,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741960_1136 (size=907469) 2025-01-08T02:20:51,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T02:20:51,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T02:20:51,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741961_1137 (size=3317408) 2025-01-08T02:20:51,438 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741962_1138 (size=503880) 2025-01-08T02:20:51,439 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741962_1138 (size=503880) 2025-01-08T02:20:51,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741962_1138 (size=503880) 2025-01-08T02:20:51,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T02:20:51,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T02:20:51,489 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741963_1139 (size=4695811) 2025-01-08T02:20:51,490 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:20:51,493 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb-testExportWithResetTtl' hfile list 2025-01-08T02:20:51,496 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=32.9 K 2025-01-08T02:20:51,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741964_1140 (size=686) 2025-01-08T02:20:51,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741964_1140 (size=686) 2025-01-08T02:20:51,530 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741964_1140 (size=686) 2025-01-08T02:20:51,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741965_1141 (size=15) 2025-01-08T02:20:51,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741965_1141 (size=15) 2025-01-08T02:20:51,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741965_1141 (size=15) 2025-01-08T02:20:52,003 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741966_1142 (size=304875) 2025-01-08T02:20:52,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741966_1142 (size=304875) 2025-01-08T02:20:52,005 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741966_1142 (size=304875) 2025-01-08T02:20:52,054 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:20:52,054 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:20:52,765 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0002_000001 (auth:SIMPLE) from 127.0.0.1:52484 2025-01-08T02:20:53,711 INFO [master/c4e5d52b9759:0.Chore.1 {}] migrate.RollingUpgradeChore(116): There is no table to migrate StoreFileTracker! 2025-01-08T02:20:53,711 INFO [master/c4e5d52b9759:0.Chore.1 {}] migrate.RollingUpgradeChore(85): All Rolling-Upgrade tasks are complete, shutdown RollingUpgradeChore! 2025-01-08T02:20:58,687 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0002_000001 (auth:SIMPLE) from 127.0.0.1:53718 2025-01-08T02:20:58,959 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741967_1143 (size=350549) 2025-01-08T02:20:58,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741967_1143 (size=350549) 2025-01-08T02:20:58,961 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741967_1143 (size=350549) 2025-01-08T02:21:01,055 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0002_000001 (auth:SIMPLE) from 127.0.0.1:60716 2025-01-08T02:21:05,306 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741968_1144 (size=14517) 2025-01-08T02:21:05,306 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741968_1144 (size=14517) 2025-01-08T02:21:05,307 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741968_1144 (size=14517) 2025-01-08T02:21:05,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741969_1145 (size=8171) 2025-01-08T02:21:05,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741969_1145 (size=8171) 2025-01-08T02:21:05,373 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741969_1145 (size=8171) 2025-01-08T02:21:05,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741970_1146 (size=5878) 2025-01-08T02:21:05,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741970_1146 (size=5878) 2025-01-08T02:21:05,422 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741970_1146 (size=5878) 2025-01-08T02:21:05,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741971_1147 (size=5102) 2025-01-08T02:21:05,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741971_1147 (size=5102) 2025-01-08T02:21:05,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741971_1147 (size=5102) 2025-01-08T02:21:05,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741972_1148 (size=17458) 2025-01-08T02:21:05,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741972_1148 (size=17458) 2025-01-08T02:21:05,588 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741972_1148 (size=17458) 2025-01-08T02:21:05,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741973_1149 (size=461) 2025-01-08T02:21:05,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741973_1149 (size=461) 2025-01-08T02:21:05,627 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741973_1149 (size=461) 2025-01-08T02:21:05,653 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_2/usercache/jenkins/appcache/application_1736302816589_0002/container_1736302816589_0002_01_000002/launch_container.sh] 2025-01-08T02:21:05,653 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_2/usercache/jenkins/appcache/application_1736302816589_0002/container_1736302816589_0002_01_000002/container_tokens] 2025-01-08T02:21:05,654 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_2/usercache/jenkins/appcache/application_1736302816589_0002/container_1736302816589_0002_01_000002/sysfs] 2025-01-08T02:21:05,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741974_1150 (size=17458) 2025-01-08T02:21:05,689 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741974_1150 (size=17458) 2025-01-08T02:21:05,691 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741974_1150 (size=17458) 2025-01-08T02:21:05,775 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741975_1151 (size=350549) 2025-01-08T02:21:05,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741975_1151 (size=350549) 2025-01-08T02:21:05,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741975_1151 (size=350549) 2025-01-08T02:21:06,389 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:21:07,216 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:21:07,217 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:21:07,225 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb-testExportWithResetTtl 2025-01-08T02:21:07,225 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:21:07,225 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:21:07,225 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb-testExportWithResetTtl at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T02:21:07,226 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb-testExportWithResetTtl/.snapshotinfo 2025-01-08T02:21:07,226 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb-testExportWithResetTtl/data.manifest 2025-01-08T02:21:07,226 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248/.hbase-snapshot/snaptb-testExportWithResetTtl at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248/.hbase-snapshot/snaptb-testExportWithResetTtl 2025-01-08T02:21:07,226 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248/.hbase-snapshot/snaptb-testExportWithResetTtl/.snapshotinfo 2025-01-08T02:21:07,226 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302849248/.hbase-snapshot/snaptb-testExportWithResetTtl/data.manifest 2025-01-08T02:21:07,233 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testExportWithResetTtl 2025-01-08T02:21:07,234 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testExportWithResetTtl 2025-01-08T02:21:07,235 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=49, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testExportWithResetTtl 2025-01-08T02:21:07,237 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T02:21:07,237 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302867237"}]},"ts":"1736302867237"} 2025-01-08T02:21:07,239 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=DISABLING in hbase:meta 2025-01-08T02:21:07,241 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testExportWithResetTtl to state=DISABLING 2025-01-08T02:21:07,242 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=50, ppid=49, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testExportWithResetTtl}] 2025-01-08T02:21:07,243 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=ebf5e7819c73f9c4f5923fe9cbc7ef7f, UNASSIGN}, {pid=52, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=7c90cc4cbd2d273c4f5ab0dde7177718, UNASSIGN}] 2025-01-08T02:21:07,244 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=52, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=7c90cc4cbd2d273c4f5ab0dde7177718, UNASSIGN 2025-01-08T02:21:07,244 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=51, ppid=50, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testExportWithResetTtl, region=ebf5e7819c73f9c4f5923fe9cbc7ef7f, UNASSIGN 2025-01-08T02:21:07,245 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=52 updating hbase:meta row=7c90cc4cbd2d273c4f5ab0dde7177718, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:21:07,245 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=ebf5e7819c73f9c4f5923fe9cbc7ef7f, regionState=CLOSING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:07,247 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:21:07,247 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=53, ppid=52, state=RUNNABLE; CloseRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:21:07,247 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:21:07,247 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=54, ppid=51, state=RUNNABLE; CloseRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:21:07,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T02:21:07,399 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:21:07,399 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:07,400 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(124): Close ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:21:07,400 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(124): Close 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1681): Closing ebf5e7819c73f9c4f5923fe9cbc7ef7f, disabling compactions & flushes 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:21:07,400 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. after waiting 0 ms 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1681): Closing 7c90cc4cbd2d273c4f5ab0dde7177718, disabling compactions & flushes 2025-01-08T02:21:07,400 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1703): Closing region testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1791): Acquired close lock on testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. after waiting 0 ms 2025-01-08T02:21:07,400 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1801): Updates disabled for region testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:21:07,406 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T02:21:07,406 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T02:21:07,407 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:21:07,407 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:21:07,407 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1922): Closed testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f. 2025-01-08T02:21:07,407 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1922): Closed testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718. 2025-01-08T02:21:07,407 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] regionserver.HRegion(1635): Region close journal for 7c90cc4cbd2d273c4f5ab0dde7177718: 2025-01-08T02:21:07,407 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] regionserver.HRegion(1635): Region close journal for ebf5e7819c73f9c4f5923fe9cbc7ef7f: 2025-01-08T02:21:07,409 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=53}] handler.UnassignRegionHandler(170): Closed 7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:21:07,410 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=52 updating hbase:meta row=7c90cc4cbd2d273c4f5ab0dde7177718, regionState=CLOSED 2025-01-08T02:21:07,410 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=54}] handler.UnassignRegionHandler(170): Closed ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:21:07,410 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=51 updating hbase:meta row=ebf5e7819c73f9c4f5923fe9cbc7ef7f, regionState=CLOSED 2025-01-08T02:21:07,413 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=53, resume processing ppid=52 2025-01-08T02:21:07,414 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=53, ppid=52, state=SUCCESS; CloseRegionProcedure 7c90cc4cbd2d273c4f5ab0dde7177718, server=c4e5d52b9759,39911,1736302808913 in 164 msec 2025-01-08T02:21:07,414 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=54, resume processing ppid=51 2025-01-08T02:21:07,414 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=54, ppid=51, state=SUCCESS; CloseRegionProcedure ebf5e7819c73f9c4f5923fe9cbc7ef7f, server=c4e5d52b9759,36743,1736302808837 in 165 msec 2025-01-08T02:21:07,415 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=52, ppid=50, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=7c90cc4cbd2d273c4f5ab0dde7177718, UNASSIGN in 170 msec 2025-01-08T02:21:07,416 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=51, resume processing ppid=50 2025-01-08T02:21:07,416 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=51, ppid=50, state=SUCCESS; TransitRegionStateProcedure table=testExportWithResetTtl, region=ebf5e7819c73f9c4f5923fe9cbc7ef7f, UNASSIGN in 171 msec 2025-01-08T02:21:07,418 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=50, resume processing ppid=49 2025-01-08T02:21:07,418 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=50, ppid=49, state=SUCCESS; CloseTableRegionsProcedure table=testExportWithResetTtl in 174 msec 2025-01-08T02:21:07,419 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302867419"}]},"ts":"1736302867419"} 2025-01-08T02:21:07,420 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportWithResetTtl, state=DISABLED in hbase:meta 2025-01-08T02:21:07,423 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testExportWithResetTtl to state=DISABLED 2025-01-08T02:21:07,424 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=49, state=SUCCESS; DisableTableProcedure table=testExportWithResetTtl in 189 msec 2025-01-08T02:21:07,539 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=49 2025-01-08T02:21:07,540 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testExportWithResetTtl, procId: 49 completed 2025-01-08T02:21:07,541 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testExportWithResetTtl 2025-01-08T02:21:07,542 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=55, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T02:21:07,543 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=55, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T02:21:07,543 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testExportWithResetTtl 2025-01-08T02:21:07,544 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=55, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T02:21:07,546 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testExportWithResetTtl 2025-01-08T02:21:07,549 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:21:07,549 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:21:07,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,552 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,553 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T02:21:07,553 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T02:21:07,553 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T02:21:07,553 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportWithResetTtl with data PBUF 2025-01-08T02:21:07,558 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/recovered.edits] 2025-01-08T02:21:07,558 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/recovered.edits] 2025-01-08T02:21:07,564 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/cf/3457da62efa340e1814230be8e434b44 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/cf/3457da62efa340e1814230be8e434b44 2025-01-08T02:21:07,564 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf/94a528e7e3454ae4849e138700b54e86 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/cf/94a528e7e3454ae4849e138700b54e86 2025-01-08T02:21:07,568 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/recovered.edits/8.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718/recovered.edits/8.seqid 2025-01-08T02:21:07,568 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/recovered.edits/8.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f/recovered.edits/8.seqid 2025-01-08T02:21:07,569 DEBUG [HFileArchiver-4 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:21:07,569 DEBUG [HFileArchiver-3 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportWithResetTtl/ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:21:07,569 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testExportWithResetTtl regions 2025-01-08T02:21:07,570 DEBUG [PEWorker-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d 2025-01-08T02:21:07,571 DEBUG [PEWorker-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf] 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testExportWithResetTtl 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:07,571 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:07,573 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:07,573 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:07,573 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:07,574 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=55 2025-01-08T02:21:07,574 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithResetTtl \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:07,579 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b202501086794d754f7d944c78f9e8e3a597f8e1e_7c90cc4cbd2d273c4f5ab0dde7177718 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/c4ca4238a0b923820dcc509a6f75849b202501086794d754f7d944c78f9e8e3a597f8e1e_7c90cc4cbd2d273c4f5ab0dde7177718 2025-01-08T02:21:07,580 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e2025010878e9ef2ff26e417a9d84db55b95e89ec_ebf5e7819c73f9c4f5923fe9cbc7ef7f to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d/cf/d41d8cd98f00b204e9800998ecf8427e2025010878e9ef2ff26e417a9d84db55b95e89ec_ebf5e7819c73f9c4f5923fe9cbc7ef7f 2025-01-08T02:21:07,581 DEBUG [PEWorker-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportWithResetTtl/ea1f2913460bad1dc9f5d962c597c09d 2025-01-08T02:21:07,584 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=55, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T02:21:07,587 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testExportWithResetTtl from hbase:meta 2025-01-08T02:21:07,589 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testExportWithResetTtl' descriptor. 2025-01-08T02:21:07,591 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=55, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T02:21:07,591 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testExportWithResetTtl' from region states. 2025-01-08T02:21:07,591 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302867591"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:07,591 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302867591"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:07,596 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:21:07,596 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => ebf5e7819c73f9c4f5923fe9cbc7ef7f, NAME => 'testExportWithResetTtl,,1736302847932.ebf5e7819c73f9c4f5923fe9cbc7ef7f.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 7c90cc4cbd2d273c4f5ab0dde7177718, NAME => 'testExportWithResetTtl,1,1736302847932.7c90cc4cbd2d273c4f5ab0dde7177718.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:21:07,596 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testExportWithResetTtl' as deleted. 2025-01-08T02:21:07,596 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302867596"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:07,599 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testExportWithResetTtl state from META 2025-01-08T02:21:07,601 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=55, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testExportWithResetTtl 2025-01-08T02:21:07,603 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=55, state=SUCCESS; DeleteTableProcedure table=testExportWithResetTtl in 60 msec 2025-01-08T02:21:07,675 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=55 2025-01-08T02:21:07,675 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testExportWithResetTtl, procId: 55 completed 2025-01-08T02:21:07,676 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithResetTtl 2025-01-08T02:21:07,677 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportWithResetTtl 2025-01-08T02:21:07,678 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=56, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:21:07,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T02:21:07,682 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302867681"}]},"ts":"1736302867681"} 2025-01-08T02:21:07,684 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=DISABLING in hbase:meta 2025-01-08T02:21:07,687 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithResetTtl to state=DISABLING 2025-01-08T02:21:07,687 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=57, ppid=56, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithResetTtl}] 2025-01-08T02:21:07,689 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=58, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=308a78ff21bb936d4a1552dc57b4435c, UNASSIGN}, {pid=59, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=922e987cbaa461a827ca105c1388a10b, UNASSIGN}] 2025-01-08T02:21:07,690 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=59, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=922e987cbaa461a827ca105c1388a10b, UNASSIGN 2025-01-08T02:21:07,690 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=58, ppid=57, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=308a78ff21bb936d4a1552dc57b4435c, UNASSIGN 2025-01-08T02:21:07,691 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=58 updating hbase:meta row=308a78ff21bb936d4a1552dc57b4435c, regionState=CLOSING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:07,691 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=59 updating hbase:meta row=922e987cbaa461a827ca105c1388a10b, regionState=CLOSING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:07,693 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:21:07,693 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=60, ppid=58, state=RUNNABLE; CloseRegionProcedure 308a78ff21bb936d4a1552dc57b4435c, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:21:07,695 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:21:07,696 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=61, ppid=59, state=RUNNABLE; CloseRegionProcedure 922e987cbaa461a827ca105c1388a10b, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:21:07,781 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T02:21:07,847 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:07,848 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(124): Close 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:21:07,848 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:21:07,848 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1681): Closing 308a78ff21bb936d4a1552dc57b4435c, disabling compactions & flushes 2025-01-08T02:21:07,849 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. after waiting 0 ms 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:21:07,849 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(124): Close 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1681): Closing 922e987cbaa461a827ca105c1388a10b, disabling compactions & flushes 2025-01-08T02:21:07,849 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1703): Closing region testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. after waiting 0 ms 2025-01-08T02:21:07,849 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:21:07,854 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:21:07,855 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:21:07,855 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c. 2025-01-08T02:21:07,855 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] regionserver.HRegion(1635): Region close journal for 308a78ff21bb936d4a1552dc57b4435c: 2025-01-08T02:21:07,857 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=60}] handler.UnassignRegionHandler(170): Closed 308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:21:07,857 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=58 updating hbase:meta row=308a78ff21bb936d4a1552dc57b4435c, regionState=CLOSED 2025-01-08T02:21:07,857 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:21:07,858 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:21:07,858 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1922): Closed testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b. 2025-01-08T02:21:07,858 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] regionserver.HRegion(1635): Region close journal for 922e987cbaa461a827ca105c1388a10b: 2025-01-08T02:21:07,860 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=61}] handler.UnassignRegionHandler(170): Closed 922e987cbaa461a827ca105c1388a10b 2025-01-08T02:21:07,861 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=59 updating hbase:meta row=922e987cbaa461a827ca105c1388a10b, regionState=CLOSED 2025-01-08T02:21:07,861 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=60, resume processing ppid=58 2025-01-08T02:21:07,861 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=60, ppid=58, state=SUCCESS; CloseRegionProcedure 308a78ff21bb936d4a1552dc57b4435c, server=c4e5d52b9759,36743,1736302808837 in 166 msec 2025-01-08T02:21:07,863 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=58, ppid=57, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=308a78ff21bb936d4a1552dc57b4435c, UNASSIGN in 172 msec 2025-01-08T02:21:07,864 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=61, resume processing ppid=59 2025-01-08T02:21:07,864 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=61, ppid=59, state=SUCCESS; CloseRegionProcedure 922e987cbaa461a827ca105c1388a10b, server=c4e5d52b9759,37031,1736302808764 in 167 msec 2025-01-08T02:21:07,865 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=59, resume processing ppid=57 2025-01-08T02:21:07,865 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=59, ppid=57, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithResetTtl, region=922e987cbaa461a827ca105c1388a10b, UNASSIGN in 175 msec 2025-01-08T02:21:07,868 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=57, resume processing ppid=56 2025-01-08T02:21:07,868 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=57, ppid=56, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithResetTtl in 179 msec 2025-01-08T02:21:07,869 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302867869"}]},"ts":"1736302867869"} 2025-01-08T02:21:07,870 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithResetTtl, state=DISABLED in hbase:meta 2025-01-08T02:21:07,873 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithResetTtl to state=DISABLED 2025-01-08T02:21:07,874 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=56, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithResetTtl in 196 msec 2025-01-08T02:21:07,983 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=56 2025-01-08T02:21:07,983 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithResetTtl, procId: 56 completed 2025-01-08T02:21:07,984 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportWithResetTtl 2025-01-08T02:21:07,985 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=62, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:21:07,986 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=62, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:21:07,986 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithResetTtl 2025-01-08T02:21:07,987 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=62, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:21:07,988 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithResetTtl 2025-01-08T02:21:07,990 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:21:07,990 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b 2025-01-08T02:21:07,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T02:21:07,992 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T02:21:07,994 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/recovered.edits] 2025-01-08T02:21:07,994 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/recovered.edits] 2025-01-08T02:21:07,997 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T02:21:07,997 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T02:21:07,997 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T02:21:07,997 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T02:21:07,997 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data PBUF 2025-01-08T02:21:07,999 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:07,999 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T02:21:07,999 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:07,999 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithResetTtl with data null 2025-01-08T02:21:07,999 INFO [zk-permission-watcher-pool-0 {}] access.AuthManager(135): Skipping permission cache refresh because writable data is empty 2025-01-08T02:21:07,999 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T02:21:07,999 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:07,999 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithResetTtl 2025-01-08T02:21:08,000 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:08,001 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=62 2025-01-08T02:21:08,001 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/cf/eeca5b5223ab4053b46a1ed3239aa3a2 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/cf/eeca5b5223ab4053b46a1ed3239aa3a2 2025-01-08T02:21:08,001 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/cf/0e0b7b1f2bb64e05a22e50c20ad43616 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/cf/0e0b7b1f2bb64e05a22e50c20ad43616 2025-01-08T02:21:08,005 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c/recovered.edits/9.seqid 2025-01-08T02:21:08,006 DEBUG [HFileArchiver-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:21:08,008 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b/recovered.edits/9.seqid 2025-01-08T02:21:08,009 DEBUG [HFileArchiver-6 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithResetTtl/922e987cbaa461a827ca105c1388a10b 2025-01-08T02:21:08,009 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithResetTtl regions 2025-01-08T02:21:08,009 DEBUG [PEWorker-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e 2025-01-08T02:21:08,011 DEBUG [PEWorker-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf] 2025-01-08T02:21:08,016 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010884eee7384bab40bcb6da79600695569b_922e987cbaa461a827ca105c1388a10b to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/c4ca4238a0b923820dcc509a6f75849b2025010884eee7384bab40bcb6da79600695569b_922e987cbaa461a827ca105c1388a10b 2025-01-08T02:21:08,018 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108683b53efe767410a96bccb9d35cd1e4f_308a78ff21bb936d4a1552dc57b4435c to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e/cf/d41d8cd98f00b204e9800998ecf8427e20250108683b53efe767410a96bccb9d35cd1e4f_308a78ff21bb936d4a1552dc57b4435c 2025-01-08T02:21:08,018 DEBUG [PEWorker-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithResetTtl/1bf72cc28ee4e494fb2ee93a7ed9fe6e 2025-01-08T02:21:08,020 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=62, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:21:08,024 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithResetTtl from hbase:meta 2025-01-08T02:21:08,026 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithResetTtl' descriptor. 2025-01-08T02:21:08,027 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=62, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:21:08,027 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithResetTtl' from region states. 2025-01-08T02:21:08,028 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302868027"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:08,028 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302868027"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:08,030 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:21:08,030 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 308a78ff21bb936d4a1552dc57b4435c, NAME => 'testtb-testExportWithResetTtl,,1736302844912.308a78ff21bb936d4a1552dc57b4435c.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 922e987cbaa461a827ca105c1388a10b, NAME => 'testtb-testExportWithResetTtl,1,1736302844912.922e987cbaa461a827ca105c1388a10b.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:21:08,030 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithResetTtl' as deleted. 2025-01-08T02:21:08,030 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithResetTtl","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302868030"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:08,032 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithResetTtl state from META 2025-01-08T02:21:08,034 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=62, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithResetTtl 2025-01-08T02:21:08,036 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=62, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithResetTtl in 50 msec 2025-01-08T02:21:08,102 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=62 2025-01-08T02:21:08,103 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithResetTtl, procId: 62 completed 2025-01-08T02:21:08,115 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportWithResetTtl" 2025-01-08T02:21:08,117 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithResetTtl 2025-01-08T02:21:08,120 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb-testExportWithResetTtl" 2025-01-08T02:21:08,123 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb-testExportWithResetTtl 2025-01-08T02:21:08,125 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportWithResetTtl" 2025-01-08T02:21:08,129 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithResetTtl 2025-01-08T02:21:08,159 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithResetTtl Thread=784 (was 768) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:45158 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:60204 [Waiting for operation #8] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_380106103_1 at /127.0.0.1:60184 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:36863 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_380106103_1 at /127.0.0.1:45140 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 29095) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-14 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:43851 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-2132 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: HFileArchiver-3 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:34178 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-16 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:36863 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-15 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-4 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:36699 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-13 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-11 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=811 (was 809) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=782 (was 815), ProcessCount=20 (was 19) - ProcessCount LEAK? -, AvailableMemoryMB=2443 (was 1516) - AvailableMemoryMB LEAK? - 2025-01-08T02:21:08,159 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=784 is superior to 500 2025-01-08T02:21:08,186 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemState Thread=784, OpenFileDescriptor=811, MaxFileDescriptor=1048576, SystemLoadAverage=782, ProcessCount=20, AvailableMemoryMB=2427 2025-01-08T02:21:08,186 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=784 is superior to 500 2025-01-08T02:21:08,188 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:21:08,192 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=63, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:08,193 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:21:08,193 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemState" procId is: 63 2025-01-08T02:21:08,194 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:21:08,195 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T02:21:08,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741976_1152 (size=443) 2025-01-08T02:21:08,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741976_1152 (size=443) 2025-01-08T02:21:08,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741976_1152 (size=443) 2025-01-08T02:21:08,233 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 5636eccc85e23586cf561b0d8ce54bc0, NAME => 'testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:08,238 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 9775d3af872866840235c4892efa4fd0, NAME => 'testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:08,266 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741977_1153 (size=68) 2025-01-08T02:21:08,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741977_1153 (size=68) 2025-01-08T02:21:08,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741977_1153 (size=68) 2025-01-08T02:21:08,268 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:08,268 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1681): Closing 5636eccc85e23586cf561b0d8ce54bc0, disabling compactions & flushes 2025-01-08T02:21:08,268 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. after waiting 0 ms 2025-01-08T02:21:08,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,269 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-0 {}] regionserver.HRegion(1635): Region close journal for 5636eccc85e23586cf561b0d8ce54bc0: 2025-01-08T02:21:08,282 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741978_1154 (size=68) 2025-01-08T02:21:08,282 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741978_1154 (size=68) 2025-01-08T02:21:08,282 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741978_1154 (size=68) 2025-01-08T02:21:08,283 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:08,283 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1681): Closing 9775d3af872866840235c4892efa4fd0, disabling compactions & flushes 2025-01-08T02:21:08,283 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:08,283 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:08,283 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. after waiting 0 ms 2025-01-08T02:21:08,283 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:08,283 INFO [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:08,283 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemState-pool-1 {}] regionserver.HRegion(1635): Region close journal for 9775d3af872866840235c4892efa4fd0: 2025-01-08T02:21:08,285 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:21:08,286 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736302868285"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302868285"}]},"ts":"1736302868285"} 2025-01-08T02:21:08,286 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736302868285"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302868285"}]},"ts":"1736302868285"} 2025-01-08T02:21:08,288 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:21:08,289 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:21:08,289 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302868289"}]},"ts":"1736302868289"} 2025-01-08T02:21:08,291 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=ENABLING in hbase:meta 2025-01-08T02:21:08,294 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:21:08,296 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T02:21:08,296 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:21:08,296 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:21:08,296 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:21:08,296 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:21:08,296 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:21:08,296 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:21:08,296 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:21:08,296 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0, ASSIGN}, {pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9775d3af872866840235c4892efa4fd0, ASSIGN}] 2025-01-08T02:21:08,297 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9775d3af872866840235c4892efa4fd0, ASSIGN 2025-01-08T02:21:08,297 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0, ASSIGN 2025-01-08T02:21:08,298 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=65, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9775d3af872866840235c4892efa4fd0, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:21:08,298 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=64, ppid=63, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:21:08,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportWithResetTtl 2025-01-08T02:21:08,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithResetTtl 2025-01-08T02:21:08,448 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:21:08,449 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=64 updating hbase:meta row=5636eccc85e23586cf561b0d8ce54bc0, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:08,449 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=65 updating hbase:meta row=9775d3af872866840235c4892efa4fd0, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:08,453 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=66, ppid=65, state=RUNNABLE; OpenRegionProcedure 9775d3af872866840235c4892efa4fd0, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:21:08,454 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=67, ppid=64, state=RUNNABLE; OpenRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:21:08,497 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T02:21:08,606 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:08,608 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:08,610 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:08,611 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7285): Opening region: {ENCODED => 9775d3af872866840235c4892efa4fd0, NAME => 'testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:21:08,611 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. service=AccessControlService 2025-01-08T02:21:08,611 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:21:08,612 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemState 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:08,612 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:08,612 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7327): checking encryption for 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:08,612 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(7330): checking classloading for 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:08,614 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,614 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7285): Opening region: {ENCODED => 5636eccc85e23586cf561b0d8ce54bc0, NAME => 'testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:21:08,614 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. service=AccessControlService 2025-01-08T02:21:08,614 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:21:08,615 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemState 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,615 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:08,615 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7327): checking encryption for 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,615 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(7330): checking classloading for 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,635 INFO [StoreOpener-9775d3af872866840235c4892efa4fd0-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:08,637 INFO [StoreOpener-9775d3af872866840235c4892efa4fd0-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 9775d3af872866840235c4892efa4fd0 columnFamilyName cf 2025-01-08T02:21:08,638 DEBUG [StoreOpener-9775d3af872866840235c4892efa4fd0-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:08,639 INFO [StoreOpener-9775d3af872866840235c4892efa4fd0-1 {}] regionserver.HStore(327): Store=9775d3af872866840235c4892efa4fd0/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:21:08,641 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:08,641 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:08,644 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1085): writing seq id for 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:08,653 INFO [StoreOpener-5636eccc85e23586cf561b0d8ce54bc0-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,656 INFO [StoreOpener-5636eccc85e23586cf561b0d8ce54bc0-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 5636eccc85e23586cf561b0d8ce54bc0 columnFamilyName cf 2025-01-08T02:21:08,657 DEBUG [StoreOpener-5636eccc85e23586cf561b0d8ce54bc0-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:08,658 INFO [StoreOpener-5636eccc85e23586cf561b0d8ce54bc0-1 {}] regionserver.HStore(327): Store=5636eccc85e23586cf561b0d8ce54bc0/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:21:08,659 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,660 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,662 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1085): writing seq id for 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,669 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:21:08,670 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1102): Opened 9775d3af872866840235c4892efa4fd0; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74800211, jitterRate=0.11461000144481659}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:21:08,671 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegion(1001): Region open journal for 9775d3af872866840235c4892efa4fd0: 2025-01-08T02:21:08,672 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0., pid=66, masterSystemTime=1736302868606 2025-01-08T02:21:08,676 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:08,676 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=66}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:08,677 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=65 updating hbase:meta row=9775d3af872866840235c4892efa4fd0, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:08,682 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=66, resume processing ppid=65 2025-01-08T02:21:08,682 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=66, ppid=65, state=SUCCESS; OpenRegionProcedure 9775d3af872866840235c4892efa4fd0, server=c4e5d52b9759,36743,1736302808837 in 227 msec 2025-01-08T02:21:08,682 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:21:08,683 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1102): Opened 5636eccc85e23586cf561b0d8ce54bc0; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68991929, jitterRate=0.028059855103492737}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:21:08,683 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegion(1001): Region open journal for 5636eccc85e23586cf561b0d8ce54bc0: 2025-01-08T02:21:08,684 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=65, ppid=63, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9775d3af872866840235c4892efa4fd0, ASSIGN in 386 msec 2025-01-08T02:21:08,684 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0., pid=67, masterSystemTime=1736302868608 2025-01-08T02:21:08,686 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,687 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=67}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,687 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=64 updating hbase:meta row=5636eccc85e23586cf561b0d8ce54bc0, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:08,690 WARN [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] assignment.AssignmentManager(1526): Unable to acquire lock for regionNode state=OPEN, location=c4e5d52b9759,37031,1736302808764, table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0. It is likely that another thread is currently holding the lock. To avoid deadlock, skip execution for now. 2025-01-08T02:21:08,694 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=67, resume processing ppid=64 2025-01-08T02:21:08,695 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=67, ppid=64, state=SUCCESS; OpenRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0, server=c4e5d52b9759,37031,1736302808764 in 238 msec 2025-01-08T02:21:08,696 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=64, resume processing ppid=63 2025-01-08T02:21:08,696 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=64, ppid=63, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0, ASSIGN in 398 msec 2025-01-08T02:21:08,697 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:21:08,698 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302868697"}]},"ts":"1736302868697"} 2025-01-08T02:21:08,699 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=ENABLED in hbase:meta 2025-01-08T02:21:08,705 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=63, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemState execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:21:08,706 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemState jenkins: RWXCA 2025-01-08T02:21:08,708 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T02:21:08,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:08,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:08,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:08,711 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:08,713 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:08,713 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:08,714 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:08,714 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:08,715 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=63, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemState in 523 msec 2025-01-08T02:21:08,799 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=63 2025-01-08T02:21:08,799 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemState, procId: 63 completed 2025-01-08T02:21:08,803 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemState 2025-01-08T02:21:08,803 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:08,803 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:21:08,828 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:21:08,828 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302868828 (current time:1736302868828). 2025-01-08T02:21:08,828 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:21:08,828 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemState VERSION not specified, setting to 2 2025-01-08T02:21:08,828 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:21:08,830 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x2a033fc4 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@2754b1ff 2025-01-08T02:21:08,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@46ef2631, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:08,887 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:08,889 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43028, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:08,891 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x2a033fc4 to 127.0.0.1:61593 2025-01-08T02:21:08,891 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:08,894 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x24d196a6 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5fd95bc8 2025-01-08T02:21:08,906 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@603820ac, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:08,912 DEBUG [hconnection-0x111d149d-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:08,914 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43036, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:08,916 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:08,917 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35206, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:08,918 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x24d196a6 to 127.0.0.1:61593 2025-01-08T02:21:08,918 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:08,919 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T02:21:08,920 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:21:08,921 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=68, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:21:08,921 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 68 2025-01-08T02:21:08,923 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:21:08,924 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T02:21:08,924 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:21:08,927 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:21:08,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741979_1155 (size=170) 2025-01-08T02:21:08,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741979_1155 (size=170) 2025-01-08T02:21:08,945 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741979_1155 (size=170) 2025-01-08T02:21:08,947 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:21:08,947 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0}, {pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0}] 2025-01-08T02:21:08,948 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:08,948 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:09,025 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T02:21:09,103 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:09,103 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:09,104 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=70 2025-01-08T02:21:09,104 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=69 2025-01-08T02:21:09,104 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:09,104 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.HRegion(2538): Flush status journal for 9775d3af872866840235c4892efa4fd0: 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.HRegion(2538): Flush status journal for 5636eccc85e23586cf561b0d8ce54bc0: 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. for emptySnaptb0-testExportFileSystemState completed. 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. for emptySnaptb0-testExportFileSystemState completed. 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.' region-info for snapshot=emptySnaptb0-testExportFileSystemState 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.' region-info for snapshot=emptySnaptb0-testExportFileSystemState 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:21:09,105 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:21:09,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741980_1156 (size=71) 2025-01-08T02:21:09,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741980_1156 (size=71) 2025-01-08T02:21:09,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741980_1156 (size=71) 2025-01-08T02:21:09,128 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:09,128 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=70}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=70 2025-01-08T02:21:09,129 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=70 2025-01-08T02:21:09,129 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemState on region 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:09,129 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=70, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:09,132 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=70, ppid=68, state=SUCCESS; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0 in 183 msec 2025-01-08T02:21:09,136 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741981_1157 (size=71) 2025-01-08T02:21:09,137 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741981_1157 (size=71) 2025-01-08T02:21:09,145 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741981_1157 (size=71) 2025-01-08T02:21:09,145 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:09,145 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=69}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=69 2025-01-08T02:21:09,146 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=69 2025-01-08T02:21:09,146 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemState on region 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:09,146 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=69, ppid=68, state=RUNNABLE; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:09,149 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=69, resume processing ppid=68 2025-01-08T02:21:09,149 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=69, ppid=68, state=SUCCESS; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0 in 200 msec 2025-01-08T02:21:09,149 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:21:09,150 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:21:09,151 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:21:09,151 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:21:09,151 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:09,152 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:21:09,163 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741982_1158 (size=63) 2025-01-08T02:21:09,163 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741982_1158 (size=63) 2025-01-08T02:21:09,168 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741982_1158 (size=63) 2025-01-08T02:21:09,169 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:21:09,170 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemState 2025-01-08T02:21:09,170 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemState 2025-01-08T02:21:09,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741983_1159 (size=653) 2025-01-08T02:21:09,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741983_1159 (size=653) 2025-01-08T02:21:09,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741983_1159 (size=653) 2025-01-08T02:21:09,215 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:21:09,221 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:21:09,221 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemState to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testExportFileSystemState 2025-01-08T02:21:09,226 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T02:21:09,226 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=68, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:21:09,227 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 68 2025-01-08T02:21:09,229 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=68, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=68, snapshot={ ss=emptySnaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } in 307 msec 2025-01-08T02:21:09,530 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=68 2025-01-08T02:21:09,531 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemState, procId: 68 completed 2025-01-08T02:21:09,544 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:21:09,554 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:21:09,563 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemState 2025-01-08T02:21:09,564 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:09,564 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:21:09,582 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:21:09,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302869583 (current time:1736302869583). 2025-01-08T02:21:09,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:21:09,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemState VERSION not specified, setting to 2 2025-01-08T02:21:09,583 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:21:09,584 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0ddfa733 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@749ff3e9 2025-01-08T02:21:09,599 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@43dcade0, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:09,601 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:09,602 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43052, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:09,604 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0ddfa733 to 127.0.0.1:61593 2025-01-08T02:21:09,604 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:09,607 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x01c1ac8d to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@a4a17e7 2025-01-08T02:21:09,620 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2c5f18e3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:09,623 DEBUG [hconnection-0x661fb4e0-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:09,624 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43066, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:09,628 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:09,629 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35218, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:09,631 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x01c1ac8d to 127.0.0.1:61593 2025-01-08T02:21:09,631 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:09,631 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T02:21:09,632 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:21:09,634 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=71, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:21:09,635 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:21:09,637 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 71 2025-01-08T02:21:09,637 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:21:09,638 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T02:21:09,640 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:21:09,665 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741984_1160 (size=165) 2025-01-08T02:21:09,667 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741984_1160 (size=165) 2025-01-08T02:21:09,667 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741984_1160 (size=165) 2025-01-08T02:21:09,668 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:21:09,669 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0}, {pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0}] 2025-01-08T02:21:09,670 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:09,671 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:09,739 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T02:21:09,823 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:09,823 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:09,823 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=72 2025-01-08T02:21:09,823 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=73 2025-01-08T02:21:09,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:09,824 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:09,824 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(2837): Flushing 5636eccc85e23586cf561b0d8ce54bc0 1/1 column families, dataSize=333 B heapSize=976 B 2025-01-08T02:21:09,824 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(2837): Flushing 9775d3af872866840235c4892efa4fd0 1/1 column families, dataSize=2.93 KB heapSize=6.58 KB 2025-01-08T02:21:09,852 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010877cb8c00c3024c3c825b3ad9f16a453f_5636eccc85e23586cf561b0d8ce54bc0 is 71, key is 03e1d16f81e6940955f7d3c471769ced/cf:q/1736302869544/Put/seqid=0 2025-01-08T02:21:09,853 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501085c36034ab04945648510cb574872770d_9775d3af872866840235c4892efa4fd0 is 71, key is 1116bd27c89b2532371d6f177fc40f19/cf:q/1736302869554/Put/seqid=0 2025-01-08T02:21:09,878 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741986_1162 (size=8031) 2025-01-08T02:21:09,879 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741985_1161 (size=5242) 2025-01-08T02:21:09,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741985_1161 (size=5242) 2025-01-08T02:21:09,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741986_1162 (size=8031) 2025-01-08T02:21:09,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741985_1161 (size=5242) 2025-01-08T02:21:09,883 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:09,883 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:09,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741986_1162 (size=8031) 2025-01-08T02:21:09,895 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b202501085c36034ab04945648510cb574872770d_9775d3af872866840235c4892efa4fd0 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085c36034ab04945648510cb574872770d_9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:09,898 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e2025010877cb8c00c3024c3c825b3ad9f16a453f_5636eccc85e23586cf561b0d8ce54bc0 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e2025010877cb8c00c3024c3c825b3ad9f16a453f_5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:09,899 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/.tmp/cf/d88911d322834ee7a00ea45e5a8800e9, store: [table=testtb-testExportFileSystemState family=cf region=9775d3af872866840235c4892efa4fd0] 2025-01-08T02:21:09,899 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/.tmp/cf/84f87cf1d0cf465e8dcaa405e8d8f12e, store: [table=testtb-testExportFileSystemState family=cf region=5636eccc85e23586cf561b0d8ce54bc0] 2025-01-08T02:21:09,900 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/.tmp/cf/d88911d322834ee7a00ea45e5a8800e9 is 209, key is 1bf2c36fc0c92408e66e96cdc2baff568/cf:q/1736302869554/Put/seqid=0 2025-01-08T02:21:09,900 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/.tmp/cf/84f87cf1d0cf465e8dcaa405e8d8f12e is 209, key is 08c4446630c764e882a4972087cf4c589/cf:q/1736302869544/Put/seqid=0 2025-01-08T02:21:09,928 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741988_1164 (size=6328) 2025-01-08T02:21:09,929 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741988_1164 (size=6328) 2025-01-08T02:21:09,929 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741988_1164 (size=6328) 2025-01-08T02:21:09,930 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=333, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/.tmp/cf/84f87cf1d0cf465e8dcaa405e8d8f12e 2025-01-08T02:21:09,938 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/.tmp/cf/84f87cf1d0cf465e8dcaa405e8d8f12e as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/cf/84f87cf1d0cf465e8dcaa405e8d8f12e 2025-01-08T02:21:09,941 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T02:21:09,944 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/cf/84f87cf1d0cf465e8dcaa405e8d8f12e, entries=5, sequenceid=6, filesize=6.2 K 2025-01-08T02:21:09,947 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(3040): Finished flush of dataSize ~333 B/333, heapSize ~960 B/960, currentSize=0 B/0 for 5636eccc85e23586cf561b0d8ce54bc0 in 122ms, sequenceid=6, compaction requested=false 2025-01-08T02:21:09,947 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemState' 2025-01-08T02:21:09,949 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.HRegion(2538): Flush status journal for 5636eccc85e23586cf561b0d8ce54bc0: 2025-01-08T02:21:09,949 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. for snaptb0-testExportFileSystemState completed. 2025-01-08T02:21:09,949 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.' region-info for snapshot=snaptb0-testExportFileSystemState 2025-01-08T02:21:09,949 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:09,949 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/cf/84f87cf1d0cf465e8dcaa405e8d8f12e] hfiles 2025-01-08T02:21:09,949 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/cf/84f87cf1d0cf465e8dcaa405e8d8f12e for snapshot=snaptb0-testExportFileSystemState 2025-01-08T02:21:09,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741987_1163 (size=14587) 2025-01-08T02:21:09,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741987_1163 (size=14587) 2025-01-08T02:21:09,956 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741987_1163 (size=14587) 2025-01-08T02:21:09,957 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=2.9 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/.tmp/cf/d88911d322834ee7a00ea45e5a8800e9 2025-01-08T02:21:09,964 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/.tmp/cf/d88911d322834ee7a00ea45e5a8800e9 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/cf/d88911d322834ee7a00ea45e5a8800e9 2025-01-08T02:21:09,972 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/cf/d88911d322834ee7a00ea45e5a8800e9, entries=45, sequenceid=6, filesize=14.2 K 2025-01-08T02:21:09,973 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(3040): Finished flush of dataSize ~2.93 KB/3003, heapSize ~6.56 KB/6720, currentSize=0 B/0 for 9775d3af872866840235c4892efa4fd0 in 149ms, sequenceid=6, compaction requested=false 2025-01-08T02:21:09,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.HRegion(2538): Flush status journal for 9775d3af872866840235c4892efa4fd0: 2025-01-08T02:21:09,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. for snaptb0-testExportFileSystemState completed. 2025-01-08T02:21:09,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.' region-info for snapshot=snaptb0-testExportFileSystemState 2025-01-08T02:21:09,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:09,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/cf/d88911d322834ee7a00ea45e5a8800e9] hfiles 2025-01-08T02:21:09,973 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/cf/d88911d322834ee7a00ea45e5a8800e9 for snapshot=snaptb0-testExportFileSystemState 2025-01-08T02:21:09,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741989_1165 (size=110) 2025-01-08T02:21:09,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741989_1165 (size=110) 2025-01-08T02:21:09,990 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741989_1165 (size=110) 2025-01-08T02:21:10,003 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741990_1166 (size=110) 2025-01-08T02:21:10,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741990_1166 (size=110) 2025-01-08T02:21:10,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741990_1166 (size=110) 2025-01-08T02:21:10,005 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:10,005 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=73}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=73 2025-01-08T02:21:10,006 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=73 2025-01-08T02:21:10,006 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemState on region 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:10,006 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=73, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:10,008 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=73, ppid=71, state=SUCCESS; SnapshotRegionProcedure 9775d3af872866840235c4892efa4fd0 in 338 msec 2025-01-08T02:21:10,242 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T02:21:10,387 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:10,387 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=72}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=72 2025-01-08T02:21:10,387 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=72 2025-01-08T02:21:10,388 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemState on region 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:10,388 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=72, ppid=71, state=RUNNABLE; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:10,390 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=72, resume processing ppid=71 2025-01-08T02:21:10,390 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=72, ppid=71, state=SUCCESS; SnapshotRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0 in 720 msec 2025-01-08T02:21:10,390 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:21:10,391 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:21:10,392 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:21:10,392 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:21:10,393 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:10,395 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085c36034ab04945648510cb574872770d_9775d3af872866840235c4892efa4fd0, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e2025010877cb8c00c3024c3c825b3ad9f16a453f_5636eccc85e23586cf561b0d8ce54bc0] hfiles 2025-01-08T02:21:10,395 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085c36034ab04945648510cb574872770d_9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:10,395 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e2025010877cb8c00c3024c3c825b3ad9f16a453f_5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:10,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741991_1167 (size=294) 2025-01-08T02:21:10,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741991_1167 (size=294) 2025-01-08T02:21:10,402 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741991_1167 (size=294) 2025-01-08T02:21:10,405 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:21:10,405 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemState 2025-01-08T02:21:10,405 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T02:21:10,418 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741992_1168 (size=963) 2025-01-08T02:21:10,418 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741992_1168 (size=963) 2025-01-08T02:21:10,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741992_1168 (size=963) 2025-01-08T02:21:10,422 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:21:10,430 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:21:10,431 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T02:21:10,432 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=71, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:21:10,432 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 71 2025-01-08T02:21:10,434 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=71, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=71, snapshot={ ss=snaptb0-testExportFileSystemState table=testtb-testExportFileSystemState type=FLUSH ttl=0 } in 800 msec 2025-01-08T02:21:10,744 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=71 2025-01-08T02:21:10,744 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemState, procId: 71 completed 2025-01-08T02:21:10,744 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744 2025-01-08T02:21:10,745 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:10,788 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:10,788 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T02:21:10,791 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:21:10,796 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemState to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemState 2025-01-08T02:21:10,824 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741993_1169 (size=165) 2025-01-08T02:21:10,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741993_1169 (size=165) 2025-01-08T02:21:10,826 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741993_1169 (size=165) 2025-01-08T02:21:10,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741994_1170 (size=963) 2025-01-08T02:21:10,839 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741994_1170 (size=963) 2025-01-08T02:21:10,841 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741994_1170 (size=963) 2025-01-08T02:21:10,843 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:10,844 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:10,844 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:10,844 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:11,487 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 4984d49fea614490cee04f171891b535 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:21:11,487 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 9775d3af872866840235c4892efa4fd0 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:21:11,488 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 8170840180427f0da936e0809aa0d2c2 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:21:11,488 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 5636eccc85e23586cf561b0d8ce54bc0 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:21:11,874 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0002_000001 (auth:SIMPLE) from 127.0.0.1:48190 2025-01-08T02:21:11,885 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0002/container_1736302816589_0002_01_000001/launch_container.sh] 2025-01-08T02:21:11,885 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0002/container_1736302816589_0002_01_000001/container_tokens] 2025-01-08T02:21:11,885 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0002/container_1736302816589_0002_01_000001/sysfs] 2025-01-08T02:21:11,956 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-9110792037921505017.jar 2025-01-08T02:21:11,957 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:11,957 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:12,027 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-14444568870641773051.jar 2025-01-08T02:21:12,027 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:12,028 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:12,028 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:12,028 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:12,029 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:12,029 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:12,029 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:21:12,029 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:21:12,030 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:21:12,030 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:21:12,030 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:21:12,030 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:21:12,031 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:21:12,031 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:21:12,031 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:21:12,031 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:21:12,031 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:21:12,032 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:21:12,032 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:12,032 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:12,032 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:12,033 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:12,033 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:12,033 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:12,034 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:12,092 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741995_1171 (size=127628) 2025-01-08T02:21:12,093 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741995_1171 (size=127628) 2025-01-08T02:21:12,093 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741995_1171 (size=127628) 2025-01-08T02:21:12,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T02:21:12,111 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T02:21:12,112 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741996_1172 (size=2172137) 2025-01-08T02:21:12,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741997_1173 (size=213228) 2025-01-08T02:21:12,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741997_1173 (size=213228) 2025-01-08T02:21:12,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741997_1173 (size=213228) 2025-01-08T02:21:12,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741998_1174 (size=1877034) 2025-01-08T02:21:12,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741998_1174 (size=1877034) 2025-01-08T02:21:12,133 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741998_1174 (size=1877034) 2025-01-08T02:21:12,141 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741999_1175 (size=533455) 2025-01-08T02:21:12,142 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741999_1175 (size=533455) 2025-01-08T02:21:12,142 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741999_1175 (size=533455) 2025-01-08T02:21:12,165 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742000_1176 (size=7280644) 2025-01-08T02:21:12,165 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742000_1176 (size=7280644) 2025-01-08T02:21:12,166 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742000_1176 (size=7280644) 2025-01-08T02:21:12,191 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742001_1177 (size=4188619) 2025-01-08T02:21:12,191 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742001_1177 (size=4188619) 2025-01-08T02:21:12,192 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742001_1177 (size=4188619) 2025-01-08T02:21:12,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742002_1178 (size=20406) 2025-01-08T02:21:12,200 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742002_1178 (size=20406) 2025-01-08T02:21:12,201 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742002_1178 (size=20406) 2025-01-08T02:21:12,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742003_1179 (size=75495) 2025-01-08T02:21:12,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742003_1179 (size=75495) 2025-01-08T02:21:12,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742003_1179 (size=75495) 2025-01-08T02:21:12,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742004_1180 (size=45609) 2025-01-08T02:21:12,216 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742004_1180 (size=45609) 2025-01-08T02:21:12,217 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742004_1180 (size=45609) 2025-01-08T02:21:12,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742005_1181 (size=110084) 2025-01-08T02:21:12,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742005_1181 (size=110084) 2025-01-08T02:21:12,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742005_1181 (size=110084) 2025-01-08T02:21:12,244 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742006_1182 (size=1323991) 2025-01-08T02:21:12,245 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742006_1182 (size=1323991) 2025-01-08T02:21:12,245 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742006_1182 (size=1323991) 2025-01-08T02:21:12,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742007_1183 (size=6350715) 2025-01-08T02:21:12,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742007_1183 (size=6350715) 2025-01-08T02:21:12,270 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742007_1183 (size=6350715) 2025-01-08T02:21:12,278 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742008_1184 (size=23076) 2025-01-08T02:21:12,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742008_1184 (size=23076) 2025-01-08T02:21:12,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742008_1184 (size=23076) 2025-01-08T02:21:12,289 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742009_1185 (size=126803) 2025-01-08T02:21:12,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742009_1185 (size=126803) 2025-01-08T02:21:12,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742009_1185 (size=126803) 2025-01-08T02:21:12,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742010_1186 (size=322274) 2025-01-08T02:21:12,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742010_1186 (size=322274) 2025-01-08T02:21:12,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742010_1186 (size=322274) 2025-01-08T02:21:12,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742011_1187 (size=1832290) 2025-01-08T02:21:12,315 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742011_1187 (size=1832290) 2025-01-08T02:21:12,316 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742011_1187 (size=1832290) 2025-01-08T02:21:12,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742012_1188 (size=30081) 2025-01-08T02:21:12,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742012_1188 (size=30081) 2025-01-08T02:21:12,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742012_1188 (size=30081) 2025-01-08T02:21:12,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742013_1189 (size=53616) 2025-01-08T02:21:12,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742013_1189 (size=53616) 2025-01-08T02:21:12,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742013_1189 (size=53616) 2025-01-08T02:21:12,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742014_1190 (size=29229) 2025-01-08T02:21:12,340 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742014_1190 (size=29229) 2025-01-08T02:21:12,341 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742014_1190 (size=29229) 2025-01-08T02:21:12,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742015_1191 (size=169089) 2025-01-08T02:21:12,350 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742015_1191 (size=169089) 2025-01-08T02:21:12,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742015_1191 (size=169089) 2025-01-08T02:21:12,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742016_1192 (size=451756) 2025-01-08T02:21:12,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742016_1192 (size=451756) 2025-01-08T02:21:12,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742016_1192 (size=451756) 2025-01-08T02:21:12,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742017_1193 (size=5175431) 2025-01-08T02:21:12,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742017_1193 (size=5175431) 2025-01-08T02:21:12,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742017_1193 (size=5175431) 2025-01-08T02:21:12,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742018_1194 (size=136454) 2025-01-08T02:21:12,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742018_1194 (size=136454) 2025-01-08T02:21:12,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742018_1194 (size=136454) 2025-01-08T02:21:12,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742019_1195 (size=907469) 2025-01-08T02:21:12,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742019_1195 (size=907469) 2025-01-08T02:21:12,419 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742019_1195 (size=907469) 2025-01-08T02:21:12,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T02:21:12,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T02:21:12,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742020_1196 (size=3317408) 2025-01-08T02:21:12,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742021_1197 (size=503880) 2025-01-08T02:21:12,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742021_1197 (size=503880) 2025-01-08T02:21:12,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742021_1197 (size=503880) 2025-01-08T02:21:12,878 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T02:21:12,878 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T02:21:12,879 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742022_1198 (size=4695811) 2025-01-08T02:21:12,880 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:21:12,883 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemState' hfile list 2025-01-08T02:21:12,885 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.4 K 2025-01-08T02:21:12,893 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742023_1199 (size=726) 2025-01-08T02:21:12,893 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742023_1199 (size=726) 2025-01-08T02:21:12,894 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742023_1199 (size=726) 2025-01-08T02:21:12,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742024_1200 (size=15) 2025-01-08T02:21:12,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742024_1200 (size=15) 2025-01-08T02:21:12,900 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742024_1200 (size=15) 2025-01-08T02:21:12,923 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742025_1201 (size=304887) 2025-01-08T02:21:12,923 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742025_1201 (size=304887) 2025-01-08T02:21:12,924 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742025_1201 (size=304887) 2025-01-08T02:21:12,948 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:21:12,948 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:21:12,952 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:21:13,884 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0003_000001 (auth:SIMPLE) from 127.0.0.1:48194 2025-01-08T02:21:18,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState 2025-01-08T02:21:18,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState Metrics about Tables on a single HBase RegionServer 2025-01-08T02:21:19,315 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0003_000001 (auth:SIMPLE) from 127.0.0.1:42596 2025-01-08T02:21:19,594 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742026_1202 (size=350561) 2025-01-08T02:21:19,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742026_1202 (size=350561) 2025-01-08T02:21:19,595 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742026_1202 (size=350561) 2025-01-08T02:21:21,562 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0003_000001 (auth:SIMPLE) from 127.0.0.1:38220 2025-01-08T02:21:24,850 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742027_1203 (size=14587) 2025-01-08T02:21:24,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742027_1203 (size=14587) 2025-01-08T02:21:24,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742027_1203 (size=14587) 2025-01-08T02:21:24,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742028_1204 (size=8031) 2025-01-08T02:21:24,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742028_1204 (size=8031) 2025-01-08T02:21:24,901 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742028_1204 (size=8031) 2025-01-08T02:21:24,949 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742029_1205 (size=6328) 2025-01-08T02:21:24,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742029_1205 (size=6328) 2025-01-08T02:21:24,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742029_1205 (size=6328) 2025-01-08T02:21:24,990 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742030_1206 (size=5242) 2025-01-08T02:21:24,991 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742030_1206 (size=5242) 2025-01-08T02:21:24,991 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742030_1206 (size=5242) 2025-01-08T02:21:25,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742031_1207 (size=17462) 2025-01-08T02:21:25,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742031_1207 (size=17462) 2025-01-08T02:21:25,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742031_1207 (size=17462) 2025-01-08T02:21:25,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742032_1208 (size=465) 2025-01-08T02:21:25,082 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742032_1208 (size=465) 2025-01-08T02:21:25,083 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742032_1208 (size=465) 2025-01-08T02:21:25,124 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742033_1209 (size=17462) 2025-01-08T02:21:25,125 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742033_1209 (size=17462) 2025-01-08T02:21:25,125 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742033_1209 (size=17462) 2025-01-08T02:21:25,139 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0003/container_1736302816589_0003_01_000002/launch_container.sh] 2025-01-08T02:21:25,139 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0003/container_1736302816589_0003_01_000002/container_tokens] 2025-01-08T02:21:25,139 WARN [ContainersLauncher #1 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0003/container_1736302816589_0003_01_000002/sysfs] 2025-01-08T02:21:25,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742034_1210 (size=350561) 2025-01-08T02:21:25,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742034_1210 (size=350561) 2025-01-08T02:21:25,147 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742034_1210 (size=350561) 2025-01-08T02:21:25,160 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0003_000001 (auth:SIMPLE) from 127.0.0.1:44568 2025-01-08T02:21:27,085 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:21:27,087 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:21:27,094 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemState 2025-01-08T02:21:27,094 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:21:27,095 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:21:27,095 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemState at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T02:21:27,095 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemState/.snapshotinfo 2025-01-08T02:21:27,095 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemState/data.manifest 2025-01-08T02:21:27,095 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744/.hbase-snapshot/snaptb0-testExportFileSystemState at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744/.hbase-snapshot/snaptb0-testExportFileSystemState 2025-01-08T02:21:27,096 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744/.hbase-snapshot/snaptb0-testExportFileSystemState/.snapshotinfo 2025-01-08T02:21:27,096 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302870744/.hbase-snapshot/snaptb0-testExportFileSystemState/data.manifest 2025-01-08T02:21:27,103 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemState 2025-01-08T02:21:27,103 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemState 2025-01-08T02:21:27,104 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=74, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:27,106 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T02:21:27,106 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302887106"}]},"ts":"1736302887106"} 2025-01-08T02:21:27,108 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=DISABLING in hbase:meta 2025-01-08T02:21:27,110 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemState to state=DISABLING 2025-01-08T02:21:27,111 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=75, ppid=74, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemState}] 2025-01-08T02:21:27,112 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=76, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0, UNASSIGN}, {pid=77, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9775d3af872866840235c4892efa4fd0, UNASSIGN}] 2025-01-08T02:21:27,112 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=77, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9775d3af872866840235c4892efa4fd0, UNASSIGN 2025-01-08T02:21:27,113 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=76, ppid=75, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0, UNASSIGN 2025-01-08T02:21:27,113 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=77 updating hbase:meta row=9775d3af872866840235c4892efa4fd0, regionState=CLOSING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:27,113 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=76 updating hbase:meta row=5636eccc85e23586cf561b0d8ce54bc0, regionState=CLOSING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:27,116 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:21:27,116 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=78, ppid=77, state=RUNNABLE; CloseRegionProcedure 9775d3af872866840235c4892efa4fd0, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:21:27,116 WARN [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063 {}] assignment.AssignmentManager(1526): Unable to acquire lock for regionNode state=CLOSING, location=c4e5d52b9759,37031,1736302808764, table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0. It is likely that another thread is currently holding the lock. To avoid deadlock, skip execution for now. 2025-01-08T02:21:27,117 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:21:27,117 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=79, ppid=76, state=RUNNABLE; CloseRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:21:27,207 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T02:21:27,269 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:27,270 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(124): Close 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:27,270 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:21:27,270 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:21:27,270 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1681): Closing 9775d3af872866840235c4892efa4fd0, disabling compactions & flushes 2025-01-08T02:21:27,270 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:27,270 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:27,270 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. after waiting 0 ms 2025-01-08T02:21:27,270 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:27,270 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(124): Close 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:27,271 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:21:27,271 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1681): Closing 5636eccc85e23586cf561b0d8ce54bc0, disabling compactions & flushes 2025-01-08T02:21:27,271 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:27,271 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:27,271 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. after waiting 0 ms 2025-01-08T02:21:27,271 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:27,276 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:21:27,276 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:21:27,277 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:21:27,277 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:21:27,277 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0. 2025-01-08T02:21:27,277 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0. 2025-01-08T02:21:27,277 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] regionserver.HRegion(1635): Region close journal for 5636eccc85e23586cf561b0d8ce54bc0: 2025-01-08T02:21:27,277 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] regionserver.HRegion(1635): Region close journal for 9775d3af872866840235c4892efa4fd0: 2025-01-08T02:21:27,278 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=78}] handler.UnassignRegionHandler(170): Closed 9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:27,283 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=77 updating hbase:meta row=9775d3af872866840235c4892efa4fd0, regionState=CLOSED 2025-01-08T02:21:27,283 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=79}] handler.UnassignRegionHandler(170): Closed 5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:27,284 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=76 updating hbase:meta row=5636eccc85e23586cf561b0d8ce54bc0, regionState=CLOSED 2025-01-08T02:21:27,287 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=78, resume processing ppid=77 2025-01-08T02:21:27,287 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=78, ppid=77, state=SUCCESS; CloseRegionProcedure 9775d3af872866840235c4892efa4fd0, server=c4e5d52b9759,36743,1736302808837 in 169 msec 2025-01-08T02:21:27,288 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=79, resume processing ppid=76 2025-01-08T02:21:27,288 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=79, ppid=76, state=SUCCESS; CloseRegionProcedure 5636eccc85e23586cf561b0d8ce54bc0, server=c4e5d52b9759,37031,1736302808764 in 169 msec 2025-01-08T02:21:27,288 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=77, ppid=75, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=9775d3af872866840235c4892efa4fd0, UNASSIGN in 175 msec 2025-01-08T02:21:27,290 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=76, resume processing ppid=75 2025-01-08T02:21:27,290 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=76, ppid=75, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemState, region=5636eccc85e23586cf561b0d8ce54bc0, UNASSIGN in 176 msec 2025-01-08T02:21:27,291 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=75, resume processing ppid=74 2025-01-08T02:21:27,291 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=75, ppid=74, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemState in 179 msec 2025-01-08T02:21:27,292 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302887292"}]},"ts":"1736302887292"} 2025-01-08T02:21:27,294 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemState, state=DISABLED in hbase:meta 2025-01-08T02:21:27,296 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemState to state=DISABLED 2025-01-08T02:21:27,298 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=74, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemState in 193 msec 2025-01-08T02:21:27,409 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=74 2025-01-08T02:21:27,409 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemState, procId: 74 completed 2025-01-08T02:21:27,410 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemState 2025-01-08T02:21:27,411 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=80, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:27,412 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=80, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:27,412 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemState 2025-01-08T02:21:27,412 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=80, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:27,414 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemState 2025-01-08T02:21:27,417 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:27,417 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,417 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,418 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,418 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,418 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T02:21:27,418 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T02:21:27,419 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T02:21:27,419 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemState with data PBUF 2025-01-08T02:21:27,419 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/recovered.edits] 2025-01-08T02:21:27,420 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:27,420 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,421 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,421 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,421 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,421 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,421 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,421 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=80 2025-01-08T02:21:27,423 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemState 2025-01-08T02:21:27,424 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,425 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/recovered.edits] 2025-01-08T02:21:27,427 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/cf/84f87cf1d0cf465e8dcaa405e8d8f12e to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/cf/84f87cf1d0cf465e8dcaa405e8d8f12e 2025-01-08T02:21:27,429 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/cf/d88911d322834ee7a00ea45e5a8800e9 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/cf/d88911d322834ee7a00ea45e5a8800e9 2025-01-08T02:21:27,430 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0/recovered.edits/9.seqid 2025-01-08T02:21:27,431 DEBUG [HFileArchiver-7 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:27,432 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0/recovered.edits/9.seqid 2025-01-08T02:21:27,433 DEBUG [HFileArchiver-8 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemState/9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:27,433 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemState regions 2025-01-08T02:21:27,433 DEBUG [PEWorker-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d 2025-01-08T02:21:27,434 DEBUG [PEWorker-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf] 2025-01-08T02:21:27,438 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085c36034ab04945648510cb574872770d_9775d3af872866840235c4892efa4fd0 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/c4ca4238a0b923820dcc509a6f75849b202501085c36034ab04945648510cb574872770d_9775d3af872866840235c4892efa4fd0 2025-01-08T02:21:27,439 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e2025010877cb8c00c3024c3c825b3ad9f16a453f_5636eccc85e23586cf561b0d8ce54bc0 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d/cf/d41d8cd98f00b204e9800998ecf8427e2025010877cb8c00c3024c3c825b3ad9f16a453f_5636eccc85e23586cf561b0d8ce54bc0 2025-01-08T02:21:27,440 DEBUG [PEWorker-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemState/8c2558c0b59bb3a4e2e578c2c8e8915d 2025-01-08T02:21:27,442 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=80, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:27,444 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemState from hbase:meta 2025-01-08T02:21:27,447 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemState' descriptor. 2025-01-08T02:21:27,447 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=80, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:27,447 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemState' from region states. 2025-01-08T02:21:27,448 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302887448"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:27,448 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302887448"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:27,450 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:21:27,450 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 5636eccc85e23586cf561b0d8ce54bc0, NAME => 'testtb-testExportFileSystemState,,1736302868188.5636eccc85e23586cf561b0d8ce54bc0.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 9775d3af872866840235c4892efa4fd0, NAME => 'testtb-testExportFileSystemState,1,1736302868188.9775d3af872866840235c4892efa4fd0.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:21:27,450 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemState' as deleted. 2025-01-08T02:21:27,450 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302887450"}]},"ts":"9223372036854775807"} 2025-01-08T02:21:27,451 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemState state from META 2025-01-08T02:21:27,454 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=80, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemState 2025-01-08T02:21:27,455 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=80, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemState in 44 msec 2025-01-08T02:21:27,522 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=80 2025-01-08T02:21:27,523 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemState, procId: 80 completed 2025-01-08T02:21:27,530 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportFileSystemState" 2025-01-08T02:21:27,532 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemState 2025-01-08T02:21:27,533 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemState" 2025-01-08T02:21:27,535 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemState 2025-01-08T02:21:27,562 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemState Thread=783 (was 784), OpenFileDescriptor=807 (was 811), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=776 (was 782), ProcessCount=14 (was 20), AvailableMemoryMB=2722 (was 2427) - AvailableMemoryMB LEAK? - 2025-01-08T02:21:27,562 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=783 is superior to 500 2025-01-08T02:21:27,581 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testConsecutiveExports Thread=783, OpenFileDescriptor=807, MaxFileDescriptor=1048576, SystemLoadAverage=776, ProcessCount=14, AvailableMemoryMB=2721 2025-01-08T02:21:27,581 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=783 is superior to 500 2025-01-08T02:21:27,583 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:21:27,584 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=81, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:21:27,585 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:21:27,585 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testConsecutiveExports" procId is: 81 2025-01-08T02:21:27,586 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:21:27,586 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T02:21:27,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742035_1211 (size=440) 2025-01-08T02:21:27,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742035_1211 (size=440) 2025-01-08T02:21:27,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742035_1211 (size=440) 2025-01-08T02:21:27,595 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 611f9566b5cd71d08f75c2bcb714d983, NAME => 'testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:27,602 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 31c0951312052282b964876c374bb1c2, NAME => 'testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testConsecutiveExports', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:27,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742036_1212 (size=65) 2025-01-08T02:21:27,611 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742036_1212 (size=65) 2025-01-08T02:21:27,612 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742036_1212 (size=65) 2025-01-08T02:21:27,613 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:27,613 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1681): Closing 611f9566b5cd71d08f75c2bcb714d983, disabling compactions & flushes 2025-01-08T02:21:27,613 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:27,613 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:27,613 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. after waiting 0 ms 2025-01-08T02:21:27,613 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:27,613 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:27,613 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-0 {}] regionserver.HRegion(1635): Region close journal for 611f9566b5cd71d08f75c2bcb714d983: 2025-01-08T02:21:27,618 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742037_1213 (size=65) 2025-01-08T02:21:27,619 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742037_1213 (size=65) 2025-01-08T02:21:27,619 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742037_1213 (size=65) 2025-01-08T02:21:27,620 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:27,620 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1681): Closing 31c0951312052282b964876c374bb1c2, disabling compactions & flushes 2025-01-08T02:21:27,620 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:27,620 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:27,620 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. after waiting 0 ms 2025-01-08T02:21:27,620 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:27,620 INFO [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:27,620 DEBUG [RegionOpenAndInit-testtb-testConsecutiveExports-pool-1 {}] regionserver.HRegion(1635): Region close journal for 31c0951312052282b964876c374bb1c2: 2025-01-08T02:21:27,621 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:21:27,622 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736302887621"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302887621"}]},"ts":"1736302887621"} 2025-01-08T02:21:27,622 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736302887621"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302887621"}]},"ts":"1736302887621"} 2025-01-08T02:21:27,624 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:21:27,625 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:21:27,625 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302887625"}]},"ts":"1736302887625"} 2025-01-08T02:21:27,626 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=ENABLING in hbase:meta 2025-01-08T02:21:27,630 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:21:27,631 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:21:27,631 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:21:27,631 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:21:27,631 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:21:27,631 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:21:27,631 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:21:27,631 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:21:27,632 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=611f9566b5cd71d08f75c2bcb714d983, ASSIGN}, {pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=31c0951312052282b964876c374bb1c2, ASSIGN}] 2025-01-08T02:21:27,632 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=31c0951312052282b964876c374bb1c2, ASSIGN 2025-01-08T02:21:27,633 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=611f9566b5cd71d08f75c2bcb714d983, ASSIGN 2025-01-08T02:21:27,633 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=83, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=31c0951312052282b964876c374bb1c2, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:21:27,633 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=82, ppid=81, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=611f9566b5cd71d08f75c2bcb714d983, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:21:27,687 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T02:21:27,784 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:21:27,784 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=83 updating hbase:meta row=31c0951312052282b964876c374bb1c2, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:21:27,784 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=82 updating hbase:meta row=611f9566b5cd71d08f75c2bcb714d983, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:27,786 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=84, ppid=82, state=RUNNABLE; OpenRegionProcedure 611f9566b5cd71d08f75c2bcb714d983, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:21:27,787 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=85, ppid=83, state=RUNNABLE; OpenRegionProcedure 31c0951312052282b964876c374bb1c2, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:21:27,889 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T02:21:27,939 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:27,941 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:21:27,942 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] handler.AssignRegionHandler(135): Open testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:27,942 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7285): Opening region: {ENCODED => 611f9566b5cd71d08f75c2bcb714d983, NAME => 'testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:21:27,943 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. service=AccessControlService 2025-01-08T02:21:27,943 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:21:27,943 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testConsecutiveExports 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:27,943 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:27,943 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7327): checking encryption for 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:27,943 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(7330): checking classloading for 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:27,945 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] handler.AssignRegionHandler(135): Open testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:27,945 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7285): Opening region: {ENCODED => 31c0951312052282b964876c374bb1c2, NAME => 'testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:21:27,945 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. service=AccessControlService 2025-01-08T02:21:27,945 INFO [StoreOpener-611f9566b5cd71d08f75c2bcb714d983-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:27,946 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:21:27,946 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testConsecutiveExports 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:27,946 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(894): Instantiated testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:21:27,946 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7327): checking encryption for 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:27,946 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(7330): checking classloading for 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:27,947 INFO [StoreOpener-611f9566b5cd71d08f75c2bcb714d983-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 611f9566b5cd71d08f75c2bcb714d983 columnFamilyName cf 2025-01-08T02:21:27,947 INFO [StoreOpener-31c0951312052282b964876c374bb1c2-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:27,948 DEBUG [StoreOpener-611f9566b5cd71d08f75c2bcb714d983-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:27,949 INFO [StoreOpener-31c0951312052282b964876c374bb1c2-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 31c0951312052282b964876c374bb1c2 columnFamilyName cf 2025-01-08T02:21:27,949 INFO [StoreOpener-611f9566b5cd71d08f75c2bcb714d983-1 {}] regionserver.HStore(327): Store=611f9566b5cd71d08f75c2bcb714d983/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:21:27,950 DEBUG [StoreOpener-31c0951312052282b964876c374bb1c2-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:27,963 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:27,963 INFO [StoreOpener-31c0951312052282b964876c374bb1c2-1 {}] regionserver.HStore(327): Store=31c0951312052282b964876c374bb1c2/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:21:27,964 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:27,964 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2 2025-01-08T02:21:27,965 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2 2025-01-08T02:21:27,966 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1085): writing seq id for 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:27,968 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1085): writing seq id for 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:27,968 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:21:27,969 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1102): Opened 611f9566b5cd71d08f75c2bcb714d983; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=70735005, jitterRate=0.054033711552619934}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:21:27,970 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegion(1001): Region open journal for 611f9566b5cd71d08f75c2bcb714d983: 2025-01-08T02:21:27,970 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:21:27,971 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1102): Opened 31c0951312052282b964876c374bb1c2; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69664283, jitterRate=0.038078710436820984}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:21:27,971 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983., pid=84, masterSystemTime=1736302887939 2025-01-08T02:21:27,971 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegion(1001): Region open journal for 31c0951312052282b964876c374bb1c2: 2025-01-08T02:21:27,972 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2., pid=85, masterSystemTime=1736302887941 2025-01-08T02:21:27,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:27,973 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=84}] handler.AssignRegionHandler(164): Opened testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:27,973 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=82 updating hbase:meta row=611f9566b5cd71d08f75c2bcb714d983, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:27,973 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:27,973 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=85}] handler.AssignRegionHandler(164): Opened testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:27,974 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=83 updating hbase:meta row=31c0951312052282b964876c374bb1c2, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:21:27,978 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=84, resume processing ppid=82 2025-01-08T02:21:27,978 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=84, ppid=82, state=SUCCESS; OpenRegionProcedure 611f9566b5cd71d08f75c2bcb714d983, server=c4e5d52b9759,36743,1736302808837 in 189 msec 2025-01-08T02:21:27,980 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=85, resume processing ppid=83 2025-01-08T02:21:27,980 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=85, ppid=83, state=SUCCESS; OpenRegionProcedure 31c0951312052282b964876c374bb1c2, server=c4e5d52b9759,39911,1736302808913 in 189 msec 2025-01-08T02:21:27,980 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=82, ppid=81, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=611f9566b5cd71d08f75c2bcb714d983, ASSIGN in 346 msec 2025-01-08T02:21:27,982 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=83, resume processing ppid=81 2025-01-08T02:21:27,982 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=83, ppid=81, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=31c0951312052282b964876c374bb1c2, ASSIGN in 348 msec 2025-01-08T02:21:27,983 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:21:27,983 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302887983"}]},"ts":"1736302887983"} 2025-01-08T02:21:27,984 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=ENABLED in hbase:meta 2025-01-08T02:21:27,988 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=81, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testConsecutiveExports execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:21:27,989 DEBUG [PEWorker-2 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testConsecutiveExports jenkins: RWXCA 2025-01-08T02:21:27,990 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T02:21:27,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,993 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,994 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:21:27,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:27,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:27,995 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:27,996 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testConsecutiveExports \x00 \x01 \x02 \x03 \x04 2025-01-08T02:21:27,997 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=81, state=SUCCESS; CreateTableProcedure table=testtb-testConsecutiveExports in 412 msec 2025-01-08T02:21:28,190 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=81 2025-01-08T02:21:28,190 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testConsecutiveExports, procId: 81 completed 2025-01-08T02:21:28,193 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testConsecutiveExports 2025-01-08T02:21:28,193 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:28,193 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:21:28,205 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T02:21:28,205 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302888205 (current time:1736302888205). 2025-01-08T02:21:28,205 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:21:28,205 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testConsecutiveExports VERSION not specified, setting to 2 2025-01-08T02:21:28,205 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:21:28,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6bca1121 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@703d04c3 2025-01-08T02:21:28,210 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@403776e9, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:28,211 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:28,212 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51374, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:28,213 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6bca1121 to 127.0.0.1:61593 2025-01-08T02:21:28,213 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:28,214 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0623ec18 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@473dd97 2025-01-08T02:21:28,216 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7f8863ce, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:28,218 DEBUG [hconnection-0x48ef8e6-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:28,219 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51390, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:28,222 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:28,222 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35400, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:28,223 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0623ec18 to 127.0.0.1:61593 2025-01-08T02:21:28,223 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:28,224 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T02:21:28,224 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:21:28,225 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=86, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T02:21:28,226 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 86 2025-01-08T02:21:28,227 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T02:21:28,227 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:21:28,228 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:21:28,231 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:21:28,246 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742038_1214 (size=161) 2025-01-08T02:21:28,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742038_1214 (size=161) 2025-01-08T02:21:28,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742038_1214 (size=161) 2025-01-08T02:21:28,249 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:21:28,250 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983}, {pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2}] 2025-01-08T02:21:28,251 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,251 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,328 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T02:21:28,402 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:21:28,402 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:28,403 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=87 2025-01-08T02:21:28,403 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=88 2025-01-08T02:21:28,403 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:28,403 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.HRegion(2538): Flush status journal for 611f9566b5cd71d08f75c2bcb714d983: 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. for emptySnaptb0-testConsecutiveExports completed. 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.' region-info for snapshot=emptySnaptb0-testConsecutiveExports 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.HRegion(2538): Flush status journal for 31c0951312052282b964876c374bb1c2: 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. for emptySnaptb0-testConsecutiveExports completed. 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.' region-info for snapshot=emptySnaptb0-testConsecutiveExports 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:28,404 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:21:28,411 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742039_1215 (size=68) 2025-01-08T02:21:28,412 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742039_1215 (size=68) 2025-01-08T02:21:28,412 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742039_1215 (size=68) 2025-01-08T02:21:28,412 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742040_1216 (size=68) 2025-01-08T02:21:28,416 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:28,416 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=87}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=87 2025-01-08T02:21:28,417 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=87 2025-01-08T02:21:28,417 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testConsecutiveExports on region 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742040_1216 (size=68) 2025-01-08T02:21:28,417 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=87, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742040_1216 (size=68) 2025-01-08T02:21:28,418 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:28,418 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=88}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=88 2025-01-08T02:21:28,418 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=88 2025-01-08T02:21:28,418 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testConsecutiveExports on region 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,422 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=88, ppid=86, state=RUNNABLE; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,423 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=87, ppid=86, state=SUCCESS; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983 in 168 msec 2025-01-08T02:21:28,424 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=88, resume processing ppid=86 2025-01-08T02:21:28,424 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:21:28,424 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=88, ppid=86, state=SUCCESS; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2 in 173 msec 2025-01-08T02:21:28,425 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:21:28,426 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:21:28,426 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:21:28,426 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:28,426 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:21:28,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports 2025-01-08T02:21:28,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports Metrics about Tables on a single HBase RegionServer 2025-01-08T02:21:28,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemState 2025-01-08T02:21:28,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742041_1217 (size=60) 2025-01-08T02:21:28,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742041_1217 (size=60) 2025-01-08T02:21:28,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742041_1217 (size=60) 2025-01-08T02:21:28,442 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:21:28,442 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testConsecutiveExports 2025-01-08T02:21:28,443 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testConsecutiveExports 2025-01-08T02:21:28,453 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742042_1218 (size=641) 2025-01-08T02:21:28,453 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742042_1218 (size=641) 2025-01-08T02:21:28,454 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742042_1218 (size=641) 2025-01-08T02:21:28,458 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:21:28,463 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:21:28,464 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testConsecutiveExports to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testConsecutiveExports 2025-01-08T02:21:28,468 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=86, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:21:28,468 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 86 2025-01-08T02:21:28,470 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=86, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=86, snapshot={ ss=emptySnaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } in 244 msec 2025-01-08T02:21:28,530 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=86 2025-01-08T02:21:28,530 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testConsecutiveExports, procId: 86 completed 2025-01-08T02:21:28,534 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.HRegion(8254): writing data to region testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:21:28,535 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:21:28,539 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testConsecutiveExports 2025-01-08T02:21:28,539 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:28,539 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:21:28,554 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T02:21:28,554 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302888554 (current time:1736302888554). 2025-01-08T02:21:28,554 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:21:28,554 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testConsecutiveExports VERSION not specified, setting to 2 2025-01-08T02:21:28,554 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:21:28,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x3bfbfb2c to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@48c17d7d 2025-01-08T02:21:28,559 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@711615a9, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:28,561 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:28,562 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51392, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:28,563 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x3bfbfb2c to 127.0.0.1:61593 2025-01-08T02:21:28,563 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:28,565 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x19fdba26 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5a9406f1 2025-01-08T02:21:28,573 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1beeeb57, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:21:28,581 DEBUG [hconnection-0x3e118eb0-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:28,582 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:51404, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:28,584 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:21:28,585 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35402, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:21:28,586 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x19fdba26 to 127.0.0.1:61593 2025-01-08T02:21:28,586 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:21:28,586 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testConsecutiveExports], kv [jenkins: RWXCA] 2025-01-08T02:21:28,587 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:21:28,588 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=89, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } 2025-01-08T02:21:28,588 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 89 2025-01-08T02:21:28,589 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T02:21:28,592 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:21:28,593 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:21:28,595 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:21:28,625 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742043_1219 (size=156) 2025-01-08T02:21:28,626 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742043_1219 (size=156) 2025-01-08T02:21:28,629 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742043_1219 (size=156) 2025-01-08T02:21:28,637 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:21:28,637 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983}, {pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2}] 2025-01-08T02:21:28,638 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,638 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,690 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T02:21:28,789 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:21:28,790 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=90 2025-01-08T02:21:28,790 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:28,791 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(2837): Flushing 611f9566b5cd71d08f75c2bcb714d983 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T02:21:28,792 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:21:28,792 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=91 2025-01-08T02:21:28,792 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:28,793 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(2837): Flushing 31c0951312052282b964876c374bb1c2 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T02:21:28,810 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108215250abda87489883041008643c41ec_611f9566b5cd71d08f75c2bcb714d983 is 71, key is 018168dc5b89ed9881a1173ebc16fc48/cf:q/1736302888533/Put/seqid=0 2025-01-08T02:21:28,826 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742044_1220 (size=5101) 2025-01-08T02:21:28,827 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742044_1220 (size=5101) 2025-01-08T02:21:28,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010857163f5044c84ca0b76593cfc5865cb7_31c0951312052282b964876c374bb1c2 is 71, key is 11a6e28c18347a4df957184d91b7e432/cf:q/1736302888535/Put/seqid=0 2025-01-08T02:21:28,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742044_1220 (size=5101) 2025-01-08T02:21:28,831 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:28,837 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108215250abda87489883041008643c41ec_611f9566b5cd71d08f75c2bcb714d983 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108215250abda87489883041008643c41ec_611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,838 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/.tmp/cf/e08f912c725448c0a817707680b32cfa, store: [table=testtb-testConsecutiveExports family=cf region=611f9566b5cd71d08f75c2bcb714d983] 2025-01-08T02:21:28,839 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/.tmp/cf/e08f912c725448c0a817707680b32cfa is 206, key is 07401304f473cc1c1976d85378527dbfa/cf:q/1736302888533/Put/seqid=0 2025-01-08T02:21:28,854 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742045_1221 (size=8172) 2025-01-08T02:21:28,855 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742045_1221 (size=8172) 2025-01-08T02:21:28,855 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742045_1221 (size=8172) 2025-01-08T02:21:28,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:28,860 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742046_1222 (size=5904) 2025-01-08T02:21:28,860 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742046_1222 (size=5904) 2025-01-08T02:21:28,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742046_1222 (size=5904) 2025-01-08T02:21:28,861 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010857163f5044c84ca0b76593cfc5865cb7_31c0951312052282b964876c374bb1c2 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b2025010857163f5044c84ca0b76593cfc5865cb7_31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,863 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/.tmp/cf/e08f912c725448c0a817707680b32cfa 2025-01-08T02:21:28,863 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/.tmp/cf/b32412062f614cf583053385b56b02aa, store: [table=testtb-testConsecutiveExports family=cf region=31c0951312052282b964876c374bb1c2] 2025-01-08T02:21:28,864 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/.tmp/cf/b32412062f614cf583053385b56b02aa is 206, key is 1077db7dc1e170fd538e49c19b76c0b54/cf:q/1736302888535/Put/seqid=0 2025-01-08T02:21:28,871 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/.tmp/cf/e08f912c725448c0a817707680b32cfa as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/cf/e08f912c725448c0a817707680b32cfa 2025-01-08T02:21:28,878 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/cf/e08f912c725448c0a817707680b32cfa, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T02:21:28,879 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for 611f9566b5cd71d08f75c2bcb714d983 in 88ms, sequenceid=6, compaction requested=false 2025-01-08T02:21:28,879 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742047_1223 (size=14855) 2025-01-08T02:21:28,879 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testConsecutiveExports' 2025-01-08T02:21:28,880 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742047_1223 (size=14855) 2025-01-08T02:21:28,881 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.HRegion(2538): Flush status journal for 611f9566b5cd71d08f75c2bcb714d983: 2025-01-08T02:21:28,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742047_1223 (size=14855) 2025-01-08T02:21:28,882 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. for snaptb0-testConsecutiveExports completed. 2025-01-08T02:21:28,882 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.' region-info for snapshot=snaptb0-testConsecutiveExports 2025-01-08T02:21:28,882 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:28,882 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/cf/e08f912c725448c0a817707680b32cfa] hfiles 2025-01-08T02:21:28,882 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/cf/e08f912c725448c0a817707680b32cfa for snapshot=snaptb0-testConsecutiveExports 2025-01-08T02:21:28,882 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/.tmp/cf/b32412062f614cf583053385b56b02aa 2025-01-08T02:21:28,889 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/.tmp/cf/b32412062f614cf583053385b56b02aa as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/cf/b32412062f614cf583053385b56b02aa 2025-01-08T02:21:28,892 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T02:21:28,895 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/cf/b32412062f614cf583053385b56b02aa, entries=47, sequenceid=6, filesize=14.5 K 2025-01-08T02:21:28,896 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for 31c0951312052282b964876c374bb1c2 in 103ms, sequenceid=6, compaction requested=false 2025-01-08T02:21:28,896 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.HRegion(2538): Flush status journal for 31c0951312052282b964876c374bb1c2: 2025-01-08T02:21:28,896 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. for snaptb0-testConsecutiveExports completed. 2025-01-08T02:21:28,897 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(241): Storing 'testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.' region-info for snapshot=snaptb0-testConsecutiveExports 2025-01-08T02:21:28,897 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:21:28,897 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/cf/b32412062f614cf583053385b56b02aa] hfiles 2025-01-08T02:21:28,897 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/cf/b32412062f614cf583053385b56b02aa for snapshot=snaptb0-testConsecutiveExports 2025-01-08T02:21:28,918 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742048_1224 (size=107) 2025-01-08T02:21:28,918 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742048_1224 (size=107) 2025-01-08T02:21:28,919 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742048_1224 (size=107) 2025-01-08T02:21:28,919 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:21:28,920 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=90}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=90 2025-01-08T02:21:28,920 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=90 2025-01-08T02:21:28,920 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testConsecutiveExports on region 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,920 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=90, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,922 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742049_1225 (size=107) 2025-01-08T02:21:28,923 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:21:28,923 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742049_1225 (size=107) 2025-01-08T02:21:28,923 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=91}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=91 2025-01-08T02:21:28,923 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=90, ppid=89, state=SUCCESS; SnapshotRegionProcedure 611f9566b5cd71d08f75c2bcb714d983 in 284 msec 2025-01-08T02:21:28,923 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742049_1225 (size=107) 2025-01-08T02:21:28,924 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=91 2025-01-08T02:21:28,924 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testConsecutiveExports on region 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,924 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=91, ppid=89, state=RUNNABLE; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,926 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=91, resume processing ppid=89 2025-01-08T02:21:28,926 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=91, ppid=89, state=SUCCESS; SnapshotRegionProcedure 31c0951312052282b964876c374bb1c2 in 287 msec 2025-01-08T02:21:28,926 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:21:28,927 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:21:28,928 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:21:28,928 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:21:28,928 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:21:28,930 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b2025010857163f5044c84ca0b76593cfc5865cb7_31c0951312052282b964876c374bb1c2, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108215250abda87489883041008643c41ec_611f9566b5cd71d08f75c2bcb714d983] hfiles 2025-01-08T02:21:28,930 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b2025010857163f5044c84ca0b76593cfc5865cb7_31c0951312052282b964876c374bb1c2 2025-01-08T02:21:28,930 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108215250abda87489883041008643c41ec_611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:21:28,942 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742050_1226 (size=291) 2025-01-08T02:21:28,942 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742050_1226 (size=291) 2025-01-08T02:21:28,942 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742050_1226 (size=291) 2025-01-08T02:21:28,944 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:21:28,944 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testConsecutiveExports 2025-01-08T02:21:28,945 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T02:21:28,964 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742051_1227 (size=951) 2025-01-08T02:21:28,964 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742051_1227 (size=951) 2025-01-08T02:21:28,965 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742051_1227 (size=951) 2025-01-08T02:21:28,970 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:21:28,976 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:21:28,977 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T02:21:28,978 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=89, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:21:28,978 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 }, snapshot procedure id = 89 2025-01-08T02:21:28,979 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=89, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=89, snapshot={ ss=snaptb0-testConsecutiveExports table=testtb-testConsecutiveExports type=FLUSH ttl=0 } in 391 msec 2025-01-08T02:21:29,193 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=89 2025-01-08T02:21:29,193 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testConsecutiveExports, procId: 89 completed 2025-01-08T02:21:29,194 INFO [Time-limited test {}] snapshot.TestExportSnapshot(476): Local export destination path: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193 2025-01-08T02:21:29,194 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:29,243 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:29,243 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@65f64b18, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T02:21:29,246 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:21:29,251 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T02:21:29,290 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:29,290 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:29,291 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:29,291 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-1711110233777134468.jar 2025-01-08T02:21:30,389 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,390 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,461 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-10499557047257371376.jar 2025-01-08T02:21:30,462 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,462 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,463 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,463 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,463 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,464 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:30,464 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:21:30,464 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:21:30,465 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:21:30,465 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:21:30,465 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:21:30,465 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:21:30,466 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:21:30,466 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:21:30,466 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:21:30,467 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:21:30,467 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:21:30,467 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:21:30,468 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:30,468 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:30,469 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:30,469 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:30,469 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:30,470 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:30,470 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:30,560 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742052_1228 (size=127628) 2025-01-08T02:21:30,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742052_1228 (size=127628) 2025-01-08T02:21:30,561 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742052_1228 (size=127628) 2025-01-08T02:21:30,573 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T02:21:30,574 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T02:21:30,574 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742053_1229 (size=2172137) 2025-01-08T02:21:30,581 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742054_1230 (size=213228) 2025-01-08T02:21:30,581 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742054_1230 (size=213228) 2025-01-08T02:21:30,582 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742054_1230 (size=213228) 2025-01-08T02:21:30,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742055_1231 (size=6350715) 2025-01-08T02:21:30,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742055_1231 (size=6350715) 2025-01-08T02:21:30,608 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742055_1231 (size=6350715) 2025-01-08T02:21:30,621 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742056_1232 (size=1877034) 2025-01-08T02:21:30,621 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742056_1232 (size=1877034) 2025-01-08T02:21:30,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742056_1232 (size=1877034) 2025-01-08T02:21:30,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742057_1233 (size=533455) 2025-01-08T02:21:30,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742057_1233 (size=533455) 2025-01-08T02:21:30,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742057_1233 (size=533455) 2025-01-08T02:21:30,673 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T02:21:30,673 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T02:21:30,674 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742058_1234 (size=7280644) 2025-01-08T02:21:30,696 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T02:21:30,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T02:21:30,697 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742059_1235 (size=4188619) 2025-01-08T02:21:30,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742060_1236 (size=20406) 2025-01-08T02:21:30,717 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742060_1236 (size=20406) 2025-01-08T02:21:30,717 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742060_1236 (size=20406) 2025-01-08T02:21:30,731 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742061_1237 (size=75495) 2025-01-08T02:21:30,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742061_1237 (size=75495) 2025-01-08T02:21:30,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742061_1237 (size=75495) 2025-01-08T02:21:30,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742062_1238 (size=45609) 2025-01-08T02:21:30,745 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742062_1238 (size=45609) 2025-01-08T02:21:30,746 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742062_1238 (size=45609) 2025-01-08T02:21:30,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742063_1239 (size=110084) 2025-01-08T02:21:30,768 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742063_1239 (size=110084) 2025-01-08T02:21:30,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742063_1239 (size=110084) 2025-01-08T02:21:30,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742064_1240 (size=1323991) 2025-01-08T02:21:30,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742064_1240 (size=1323991) 2025-01-08T02:21:30,786 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742064_1240 (size=1323991) 2025-01-08T02:21:30,802 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742065_1241 (size=23076) 2025-01-08T02:21:30,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742065_1241 (size=23076) 2025-01-08T02:21:30,803 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742065_1241 (size=23076) 2025-01-08T02:21:30,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742066_1242 (size=126803) 2025-01-08T02:21:30,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742066_1242 (size=126803) 2025-01-08T02:21:30,813 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742066_1242 (size=126803) 2025-01-08T02:21:30,822 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742067_1243 (size=322274) 2025-01-08T02:21:30,822 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742067_1243 (size=322274) 2025-01-08T02:21:30,823 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742067_1243 (size=322274) 2025-01-08T02:21:30,839 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742068_1244 (size=1832290) 2025-01-08T02:21:30,839 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742068_1244 (size=1832290) 2025-01-08T02:21:30,839 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742068_1244 (size=1832290) 2025-01-08T02:21:31,250 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742069_1245 (size=30081) 2025-01-08T02:21:31,251 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742069_1245 (size=30081) 2025-01-08T02:21:31,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742069_1245 (size=30081) 2025-01-08T02:21:31,267 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0003_000001 (auth:SIMPLE) from 127.0.0.1:44578 2025-01-08T02:21:31,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742070_1246 (size=451756) 2025-01-08T02:21:31,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742070_1246 (size=451756) 2025-01-08T02:21:31,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742070_1246 (size=451756) 2025-01-08T02:21:31,289 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_3/usercache/jenkins/appcache/application_1736302816589_0003/container_1736302816589_0003_01_000001/launch_container.sh] 2025-01-08T02:21:31,290 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_3/usercache/jenkins/appcache/application_1736302816589_0003/container_1736302816589_0003_01_000001/container_tokens] 2025-01-08T02:21:31,290 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_3/usercache/jenkins/appcache/application_1736302816589_0003/container_1736302816589_0003_01_000001/sysfs] 2025-01-08T02:21:31,302 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742071_1247 (size=53616) 2025-01-08T02:21:31,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742071_1247 (size=53616) 2025-01-08T02:21:31,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742071_1247 (size=53616) 2025-01-08T02:21:31,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742072_1248 (size=29229) 2025-01-08T02:21:31,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742072_1248 (size=29229) 2025-01-08T02:21:31,319 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742072_1248 (size=29229) 2025-01-08T02:21:31,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742073_1249 (size=169089) 2025-01-08T02:21:31,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742073_1249 (size=169089) 2025-01-08T02:21:31,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742073_1249 (size=169089) 2025-01-08T02:21:31,798 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742074_1250 (size=5175431) 2025-01-08T02:21:31,798 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742074_1250 (size=5175431) 2025-01-08T02:21:31,798 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742074_1250 (size=5175431) 2025-01-08T02:21:31,810 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742075_1251 (size=136454) 2025-01-08T02:21:31,811 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742075_1251 (size=136454) 2025-01-08T02:21:31,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742075_1251 (size=136454) 2025-01-08T02:21:31,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742076_1252 (size=907469) 2025-01-08T02:21:31,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742076_1252 (size=907469) 2025-01-08T02:21:31,825 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742076_1252 (size=907469) 2025-01-08T02:21:31,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T02:21:31,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T02:21:31,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742077_1253 (size=3317408) 2025-01-08T02:21:31,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742078_1254 (size=503880) 2025-01-08T02:21:31,882 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742078_1254 (size=503880) 2025-01-08T02:21:31,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742078_1254 (size=503880) 2025-01-08T02:21:31,914 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T02:21:31,914 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T02:21:31,915 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742079_1255 (size=4695811) 2025-01-08T02:21:31,916 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:21:31,920 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testConsecutiveExports' hfile list 2025-01-08T02:21:31,922 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T02:21:31,942 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742080_1256 (size=714) 2025-01-08T02:21:31,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742080_1256 (size=714) 2025-01-08T02:21:31,945 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742080_1256 (size=714) 2025-01-08T02:21:31,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742081_1257 (size=15) 2025-01-08T02:21:31,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742081_1257 (size=15) 2025-01-08T02:21:31,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742081_1257 (size=15) 2025-01-08T02:21:31,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742082_1258 (size=304928) 2025-01-08T02:21:31,997 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742082_1258 (size=304928) 2025-01-08T02:21:31,998 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742082_1258 (size=304928) 2025-01-08T02:21:32,020 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:21:32,021 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:21:32,268 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0004_000001 (auth:SIMPLE) from 127.0.0.1:44586 2025-01-08T02:21:32,812 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:21:36,389 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:21:38,869 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0004_000001 (auth:SIMPLE) from 127.0.0.1:38068 2025-01-08T02:21:39,127 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742083_1259 (size=350602) 2025-01-08T02:21:39,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742083_1259 (size=350602) 2025-01-08T02:21:39,128 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742083_1259 (size=350602) 2025-01-08T02:21:41,155 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0004_000001 (auth:SIMPLE) from 127.0.0.1:34004 2025-01-08T02:21:45,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742084_1260 (size=17451) 2025-01-08T02:21:45,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742084_1260 (size=17451) 2025-01-08T02:21:45,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742084_1260 (size=17451) 2025-01-08T02:21:45,426 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742085_1261 (size=462) 2025-01-08T02:21:45,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742085_1261 (size=462) 2025-01-08T02:21:45,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742085_1261 (size=462) 2025-01-08T02:21:45,479 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742086_1262 (size=17451) 2025-01-08T02:21:45,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742086_1262 (size=17451) 2025-01-08T02:21:45,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742086_1262 (size=17451) 2025-01-08T02:21:45,504 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0004/container_1736302816589_0004_01_000002/launch_container.sh] 2025-01-08T02:21:45,504 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0004/container_1736302816589_0004_01_000002/container_tokens] 2025-01-08T02:21:45,504 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0004/container_1736302816589_0004_01_000002/sysfs] 2025-01-08T02:21:45,508 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742087_1263 (size=350602) 2025-01-08T02:21:45,508 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742087_1263 (size=350602) 2025-01-08T02:21:45,509 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742087_1263 (size=350602) 2025-01-08T02:21:45,526 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0004_000001 (auth:SIMPLE) from 127.0.0.1:33792 2025-01-08T02:21:47,200 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:21:47,200 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:21:47,204 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testConsecutiveExports 2025-01-08T02:21:47,204 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:21:47,204 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:21:47,204 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T02:21:47,209 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T02:21:47,209 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T02:21:47,209 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in org.apache.hadoop.fs.LocalFileSystem@65f64b18 in root file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports at file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T02:21:47,209 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T02:21:47,209 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T02:21:47,211 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:47,250 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:21:47,250 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@65f64b18, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T02:21:47,253 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:21:47,259 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/.tmp/snaptb0-testConsecutiveExports 2025-01-08T02:21:47,283 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:47,283 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:47,284 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:47,284 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,457 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-1232576705184992171.jar 2025-01-08T02:21:48,458 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,458 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,533 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-13772766730705287252.jar 2025-01-08T02:21:48,533 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,534 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:21:48,535 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:21:48,535 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:21:48,535 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:21:48,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:21:48,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:21:48,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:21:48,537 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:21:48,537 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:21:48,537 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:21:48,537 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:21:48,538 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:21:48,538 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:21:48,538 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:48,538 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:48,539 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:48,539 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:48,539 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:21:48,539 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:48,539 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:21:48,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742088_1264 (size=127628) 2025-01-08T02:21:48,596 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742088_1264 (size=127628) 2025-01-08T02:21:48,597 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742088_1264 (size=127628) 2025-01-08T02:21:48,618 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T02:21:48,619 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T02:21:48,619 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742089_1265 (size=2172137) 2025-01-08T02:21:48,634 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742090_1266 (size=213228) 2025-01-08T02:21:48,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742090_1266 (size=213228) 2025-01-08T02:21:48,635 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742090_1266 (size=213228) 2025-01-08T02:21:48,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T02:21:48,652 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T02:21:48,653 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742091_1267 (size=1877034) 2025-01-08T02:21:48,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742092_1268 (size=533455) 2025-01-08T02:21:48,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742092_1268 (size=533455) 2025-01-08T02:21:48,662 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742092_1268 (size=533455) 2025-01-08T02:21:48,694 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742093_1269 (size=6350715) 2025-01-08T02:21:48,694 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742093_1269 (size=6350715) 2025-01-08T02:21:48,694 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742093_1269 (size=6350715) 2025-01-08T02:21:48,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T02:21:48,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T02:21:48,738 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742094_1270 (size=7280644) 2025-01-08T02:21:48,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742095_1271 (size=4188619) 2025-01-08T02:21:48,767 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742095_1271 (size=4188619) 2025-01-08T02:21:48,767 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742095_1271 (size=4188619) 2025-01-08T02:21:48,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742096_1272 (size=20406) 2025-01-08T02:21:48,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742096_1272 (size=20406) 2025-01-08T02:21:48,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742096_1272 (size=20406) 2025-01-08T02:21:48,787 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742097_1273 (size=75495) 2025-01-08T02:21:48,787 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742097_1273 (size=75495) 2025-01-08T02:21:48,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742097_1273 (size=75495) 2025-01-08T02:21:48,796 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742098_1274 (size=45609) 2025-01-08T02:21:48,796 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742098_1274 (size=45609) 2025-01-08T02:21:48,796 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742098_1274 (size=45609) 2025-01-08T02:21:48,812 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742099_1275 (size=110084) 2025-01-08T02:21:48,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742099_1275 (size=110084) 2025-01-08T02:21:48,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742099_1275 (size=110084) 2025-01-08T02:21:48,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742100_1276 (size=1323991) 2025-01-08T02:21:48,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742100_1276 (size=1323991) 2025-01-08T02:21:48,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742100_1276 (size=1323991) 2025-01-08T02:21:48,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742101_1277 (size=23076) 2025-01-08T02:21:48,844 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742101_1277 (size=23076) 2025-01-08T02:21:48,845 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742101_1277 (size=23076) 2025-01-08T02:21:48,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742102_1278 (size=126803) 2025-01-08T02:21:48,861 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742102_1278 (size=126803) 2025-01-08T02:21:48,862 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742102_1278 (size=126803) 2025-01-08T02:21:48,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742103_1279 (size=322274) 2025-01-08T02:21:48,871 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742103_1279 (size=322274) 2025-01-08T02:21:48,872 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742103_1279 (size=322274) 2025-01-08T02:21:48,888 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742104_1280 (size=1832290) 2025-01-08T02:21:48,888 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742104_1280 (size=1832290) 2025-01-08T02:21:48,889 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742104_1280 (size=1832290) 2025-01-08T02:21:48,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742105_1281 (size=30081) 2025-01-08T02:21:48,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742105_1281 (size=30081) 2025-01-08T02:21:48,932 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742105_1281 (size=30081) 2025-01-08T02:21:48,964 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742106_1282 (size=53616) 2025-01-08T02:21:48,964 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742106_1282 (size=53616) 2025-01-08T02:21:48,965 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742106_1282 (size=53616) 2025-01-08T02:21:49,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742107_1283 (size=29229) 2025-01-08T02:21:49,017 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742107_1283 (size=29229) 2025-01-08T02:21:49,018 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742107_1283 (size=29229) 2025-01-08T02:21:49,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742108_1284 (size=169089) 2025-01-08T02:21:49,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742108_1284 (size=169089) 2025-01-08T02:21:49,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742108_1284 (size=169089) 2025-01-08T02:21:49,058 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742109_1285 (size=451756) 2025-01-08T02:21:49,058 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742109_1285 (size=451756) 2025-01-08T02:21:49,059 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742109_1285 (size=451756) 2025-01-08T02:21:49,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742110_1286 (size=5175431) 2025-01-08T02:21:49,086 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742110_1286 (size=5175431) 2025-01-08T02:21:49,087 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742110_1286 (size=5175431) 2025-01-08T02:21:49,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742111_1287 (size=136454) 2025-01-08T02:21:49,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742111_1287 (size=136454) 2025-01-08T02:21:49,099 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742111_1287 (size=136454) 2025-01-08T02:21:49,117 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742112_1288 (size=907469) 2025-01-08T02:21:49,117 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742112_1288 (size=907469) 2025-01-08T02:21:49,118 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742112_1288 (size=907469) 2025-01-08T02:21:49,139 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T02:21:49,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T02:21:49,140 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742113_1289 (size=3317408) 2025-01-08T02:21:49,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742114_1290 (size=503880) 2025-01-08T02:21:49,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742114_1290 (size=503880) 2025-01-08T02:21:49,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742114_1290 (size=503880) 2025-01-08T02:21:49,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T02:21:49,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T02:21:49,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742115_1291 (size=4695811) 2025-01-08T02:21:49,176 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:21:49,178 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testConsecutiveExports' hfile list 2025-01-08T02:21:49,180 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T02:21:49,187 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742116_1292 (size=714) 2025-01-08T02:21:49,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742116_1292 (size=714) 2025-01-08T02:21:49,188 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742116_1292 (size=714) 2025-01-08T02:21:49,208 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742117_1293 (size=15) 2025-01-08T02:21:49,209 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742117_1293 (size=15) 2025-01-08T02:21:49,212 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742117_1293 (size=15) 2025-01-08T02:21:49,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742118_1294 (size=304928) 2025-01-08T02:21:49,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742118_1294 (size=304928) 2025-01-08T02:21:49,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742118_1294 (size=304928) 2025-01-08T02:21:51,610 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:21:51,610 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:21:51,613 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0004_000001 (auth:SIMPLE) from 127.0.0.1:52146 2025-01-08T02:21:51,624 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0004/container_1736302816589_0004_01_000001/launch_container.sh] 2025-01-08T02:21:51,624 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0004/container_1736302816589_0004_01_000001/container_tokens] 2025-01-08T02:21:51,624 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_0/usercache/jenkins/appcache/application_1736302816589_0004/container_1736302816589_0004_01_000001/sysfs] 2025-01-08T02:21:52,486 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0005_000001 (auth:SIMPLE) from 127.0.0.1:33808 2025-01-08T02:21:58,691 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0005_000001 (auth:SIMPLE) from 127.0.0.1:51906 2025-01-08T02:21:58,954 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742119_1295 (size=350602) 2025-01-08T02:21:58,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742119_1295 (size=350602) 2025-01-08T02:21:58,955 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742119_1295 (size=350602) 2025-01-08T02:22:01,005 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0005_000001 (auth:SIMPLE) from 127.0.0.1:59156 2025-01-08T02:22:04,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742120_1296 (size=16925) 2025-01-08T02:22:04,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742120_1296 (size=16925) 2025-01-08T02:22:04,279 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742120_1296 (size=16925) 2025-01-08T02:22:04,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742121_1297 (size=462) 2025-01-08T02:22:04,298 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742121_1297 (size=462) 2025-01-08T02:22:04,299 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742121_1297 (size=462) 2025-01-08T02:22:04,320 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0005/container_1736302816589_0005_01_000002/launch_container.sh] 2025-01-08T02:22:04,320 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0005/container_1736302816589_0005_01_000002/container_tokens] 2025-01-08T02:22:04,320 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0005/container_1736302816589_0005_01_000002/sysfs] 2025-01-08T02:22:04,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742122_1298 (size=16925) 2025-01-08T02:22:04,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742122_1298 (size=16925) 2025-01-08T02:22:04,343 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742122_1298 (size=16925) 2025-01-08T02:22:04,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742123_1299 (size=350602) 2025-01-08T02:22:04,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742123_1299 (size=350602) 2025-01-08T02:22:04,365 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742123_1299 (size=350602) 2025-01-08T02:22:04,381 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0005_000001 (auth:SIMPLE) from 127.0.0.1:59162 2025-01-08T02:22:05,437 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:22:05,438 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:22:05,453 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testConsecutiveExports 2025-01-08T02:22:05,453 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:22:05,453 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:22:05,453 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T02:22:05,457 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T02:22:05,457 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T02:22:05,457 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in org.apache.hadoop.fs.LocalFileSystem@65f64b18 in root file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports at file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports 2025-01-08T02:22:05,458 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports/data.manifest 2025-01-08T02:22:05,458 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302889193/.hbase-snapshot/snaptb0-testConsecutiveExports/.snapshotinfo 2025-01-08T02:22:05,482 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testConsecutiveExports 2025-01-08T02:22:05,483 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testConsecutiveExports 2025-01-08T02:22:05,484 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=92, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:22:05,487 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T02:22:05,489 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302925489"}]},"ts":"1736302925489"} 2025-01-08T02:22:05,491 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=DISABLING in hbase:meta 2025-01-08T02:22:05,493 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(284): Set testtb-testConsecutiveExports to state=DISABLING 2025-01-08T02:22:05,494 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=93, ppid=92, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testConsecutiveExports}] 2025-01-08T02:22:05,496 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=94, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=611f9566b5cd71d08f75c2bcb714d983, UNASSIGN}, {pid=95, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=31c0951312052282b964876c374bb1c2, UNASSIGN}] 2025-01-08T02:22:05,497 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=95, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=31c0951312052282b964876c374bb1c2, UNASSIGN 2025-01-08T02:22:05,497 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=94, ppid=93, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=611f9566b5cd71d08f75c2bcb714d983, UNASSIGN 2025-01-08T02:22:05,498 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=94 updating hbase:meta row=611f9566b5cd71d08f75c2bcb714d983, regionState=CLOSING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:05,498 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=95 updating hbase:meta row=31c0951312052282b964876c374bb1c2, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:05,500 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:05,500 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=96, ppid=95, state=RUNNABLE; CloseRegionProcedure 31c0951312052282b964876c374bb1c2, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:05,501 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:05,501 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=97, ppid=94, state=RUNNABLE; CloseRegionProcedure 611f9566b5cd71d08f75c2bcb714d983, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:22:05,588 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T02:22:05,652 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:05,652 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:05,653 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(124): Close 31c0951312052282b964876c374bb1c2 2025-01-08T02:22:05,653 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:05,653 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1681): Closing 31c0951312052282b964876c374bb1c2, disabling compactions & flushes 2025-01-08T02:22:05,653 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(124): Close 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:22:05,653 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:22:05,653 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:05,654 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:22:05,654 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. after waiting 0 ms 2025-01-08T02:22:05,654 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:22:05,654 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1681): Closing 611f9566b5cd71d08f75c2bcb714d983, disabling compactions & flushes 2025-01-08T02:22:05,654 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1703): Closing region testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:22:05,654 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:22:05,654 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1791): Acquired close lock on testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. after waiting 0 ms 2025-01-08T02:22:05,654 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1801): Updates disabled for region testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:22:05,685 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:05,686 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:05,687 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983. 2025-01-08T02:22:05,687 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] regionserver.HRegion(1635): Region close journal for 611f9566b5cd71d08f75c2bcb714d983: 2025-01-08T02:22:05,689 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=97}] handler.UnassignRegionHandler(170): Closed 611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:22:05,689 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=94 updating hbase:meta row=611f9566b5cd71d08f75c2bcb714d983, regionState=CLOSED 2025-01-08T02:22:05,693 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=97, resume processing ppid=94 2025-01-08T02:22:05,693 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=97, ppid=94, state=SUCCESS; CloseRegionProcedure 611f9566b5cd71d08f75c2bcb714d983, server=c4e5d52b9759,36743,1736302808837 in 190 msec 2025-01-08T02:22:05,695 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=94, ppid=93, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=611f9566b5cd71d08f75c2bcb714d983, UNASSIGN in 197 msec 2025-01-08T02:22:05,697 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:05,698 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:05,698 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1922): Closed testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2. 2025-01-08T02:22:05,698 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] regionserver.HRegion(1635): Region close journal for 31c0951312052282b964876c374bb1c2: 2025-01-08T02:22:05,699 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=96}] handler.UnassignRegionHandler(170): Closed 31c0951312052282b964876c374bb1c2 2025-01-08T02:22:05,700 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=95 updating hbase:meta row=31c0951312052282b964876c374bb1c2, regionState=CLOSED 2025-01-08T02:22:05,705 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=96, resume processing ppid=95 2025-01-08T02:22:05,705 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=96, ppid=95, state=SUCCESS; CloseRegionProcedure 31c0951312052282b964876c374bb1c2, server=c4e5d52b9759,39911,1736302808913 in 202 msec 2025-01-08T02:22:05,711 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=95, resume processing ppid=93 2025-01-08T02:22:05,711 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=95, ppid=93, state=SUCCESS; TransitRegionStateProcedure table=testtb-testConsecutiveExports, region=31c0951312052282b964876c374bb1c2, UNASSIGN in 209 msec 2025-01-08T02:22:05,716 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=93, resume processing ppid=92 2025-01-08T02:22:05,716 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=93, ppid=92, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testConsecutiveExports in 219 msec 2025-01-08T02:22:05,718 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302925718"}]},"ts":"1736302925718"} 2025-01-08T02:22:05,721 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testConsecutiveExports, state=DISABLED in hbase:meta 2025-01-08T02:22:05,734 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(296): Set testtb-testConsecutiveExports to state=DISABLED 2025-01-08T02:22:05,737 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=92, state=SUCCESS; DisableTableProcedure table=testtb-testConsecutiveExports in 252 msec 2025-01-08T02:22:05,789 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=92 2025-01-08T02:22:05,789 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testConsecutiveExports, procId: 92 completed 2025-01-08T02:22:05,790 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testConsecutiveExports 2025-01-08T02:22:05,791 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=98, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:22:05,792 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=98, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:22:05,792 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testConsecutiveExports 2025-01-08T02:22:05,792 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=98, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:22:05,794 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testConsecutiveExports 2025-01-08T02:22:05,796 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:22:05,796 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2 2025-01-08T02:22:05,798 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/recovered.edits] 2025-01-08T02:22:05,798 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/recovered.edits] 2025-01-08T02:22:05,803 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/cf/e08f912c725448c0a817707680b32cfa to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/cf/e08f912c725448c0a817707680b32cfa 2025-01-08T02:22:05,803 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/cf/b32412062f614cf583053385b56b02aa to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/cf/b32412062f614cf583053385b56b02aa 2025-01-08T02:22:05,806 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983/recovered.edits/9.seqid 2025-01-08T02:22:05,807 DEBUG [HFileArchiver-9 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:22:05,807 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2/recovered.edits/9.seqid 2025-01-08T02:22:05,808 DEBUG [HFileArchiver-10 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testConsecutiveExports/31c0951312052282b964876c374bb1c2 2025-01-08T02:22:05,808 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testConsecutiveExports regions 2025-01-08T02:22:05,808 DEBUG [PEWorker-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac 2025-01-08T02:22:05,809 DEBUG [PEWorker-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf] 2025-01-08T02:22:05,813 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b2025010857163f5044c84ca0b76593cfc5865cb7_31c0951312052282b964876c374bb1c2 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/c4ca4238a0b923820dcc509a6f75849b2025010857163f5044c84ca0b76593cfc5865cb7_31c0951312052282b964876c374bb1c2 2025-01-08T02:22:05,815 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108215250abda87489883041008643c41ec_611f9566b5cd71d08f75c2bcb714d983 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac/cf/d41d8cd98f00b204e9800998ecf8427e20250108215250abda87489883041008643c41ec_611f9566b5cd71d08f75c2bcb714d983 2025-01-08T02:22:05,815 DEBUG [PEWorker-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testConsecutiveExports/c7b48a2e62736c517f8bfa7d64fc37ac 2025-01-08T02:22:05,818 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=98, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:22:05,821 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testConsecutiveExports from hbase:meta 2025-01-08T02:22:05,827 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testConsecutiveExports' descriptor. 2025-01-08T02:22:05,833 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=98, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:22:05,833 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testConsecutiveExports' from region states. 2025-01-08T02:22:05,833 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302925833"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:05,834 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302925833"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:05,836 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:22:05,836 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 611f9566b5cd71d08f75c2bcb714d983, NAME => 'testtb-testConsecutiveExports,,1736302887582.611f9566b5cd71d08f75c2bcb714d983.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 31c0951312052282b964876c374bb1c2, NAME => 'testtb-testConsecutiveExports,1,1736302887582.31c0951312052282b964876c374bb1c2.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:22:05,836 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testConsecutiveExports' as deleted. 2025-01-08T02:22:05,836 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testConsecutiveExports","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302925836"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:05,838 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testConsecutiveExports state from META 2025-01-08T02:22:05,857 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,857 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,857 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,857 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,896 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T02:22:05,896 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T02:22:05,896 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T02:22:05,896 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testConsecutiveExports with data PBUF 2025-01-08T02:22:05,896 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=98, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testConsecutiveExports 2025-01-08T02:22:05,898 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=98, state=SUCCESS; DeleteTableProcedure table=testtb-testConsecutiveExports in 106 msec 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testConsecutiveExports 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:05,903 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:05,904 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=98 2025-01-08T02:22:05,904 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testConsecutiveExports, procId: 98 completed 2025-01-08T02:22:05,912 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testConsecutiveExports" 2025-01-08T02:22:05,914 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testConsecutiveExports 2025-01-08T02:22:05,915 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testConsecutiveExports" 2025-01-08T02:22:05,916 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testConsecutiveExports 2025-01-08T02:22:05,937 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testConsecutiveExports Thread=784 (was 783) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-25 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:55070 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #7 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ContainersLauncher #2 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-10 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-22 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:43693 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: ApplicationMasterLauncher #6 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-9 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 4901) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1526253590_1 at /127.0.0.1:48908 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-4092 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:43693 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-24 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #5 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-23 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: LogDeleter #1 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1177) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:54694 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #8 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_1526253590_1 at /127.0.0.1:35872 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=803 (was 807), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=780 (was 776) - SystemLoadAverage LEAK? -, ProcessCount=20 (was 14) - ProcessCount LEAK? -, AvailableMemoryMB=2040 (was 2721) 2025-01-08T02:22:05,937 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=784 is superior to 500 2025-01-08T02:22:05,959 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithMergeRegion Thread=784, OpenFileDescriptor=803, MaxFileDescriptor=1048576, SystemLoadAverage=780, ProcessCount=20, AvailableMemoryMB=2039 2025-01-08T02:22:05,959 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=784 is superior to 500 2025-01-08T02:22:05,961 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:22:05,962 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=99, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:05,963 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:22:05,963 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithMergeRegion" procId is: 99 2025-01-08T02:22:05,964 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:22:05,964 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T02:22:05,982 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742124_1300 (size=458) 2025-01-08T02:22:05,982 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742124_1300 (size=458) 2025-01-08T02:22:05,982 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742124_1300 (size=458) 2025-01-08T02:22:05,984 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 1a942e02b6f510a0af53f37e83ed048b, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:05,984 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => b0d1552d5ae3c896125d7ba16ec8a867, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:06,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742125_1301 (size=83) 2025-01-08T02:22:06,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742125_1301 (size=83) 2025-01-08T02:22:06,000 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742125_1301 (size=83) 2025-01-08T02:22:06,001 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:06,001 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1681): Closing b0d1552d5ae3c896125d7ba16ec8a867, disabling compactions & flushes 2025-01-08T02:22:06,001 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,001 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,001 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. after waiting 0 ms 2025-01-08T02:22:06,001 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,001 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,001 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-1 {}] regionserver.HRegion(1635): Region close journal for b0d1552d5ae3c896125d7ba16ec8a867: 2025-01-08T02:22:06,002 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742126_1302 (size=83) 2025-01-08T02:22:06,002 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742126_1302 (size=83) 2025-01-08T02:22:06,003 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742126_1302 (size=83) 2025-01-08T02:22:06,003 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:06,004 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1681): Closing 1a942e02b6f510a0af53f37e83ed048b, disabling compactions & flushes 2025-01-08T02:22:06,004 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,004 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,004 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. after waiting 0 ms 2025-01-08T02:22:06,004 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,004 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,004 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-pool-0 {}] regionserver.HRegion(1635): Region close journal for 1a942e02b6f510a0af53f37e83ed048b: 2025-01-08T02:22:06,005 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:22:06,005 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.","families":{"info":[{"qualifier":"regioninfo","vlen":82,"tag":[],"timestamp":"1736302926005"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302926005"}]},"ts":"1736302926005"} 2025-01-08T02:22:06,005 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.","families":{"info":[{"qualifier":"regioninfo","vlen":82,"tag":[],"timestamp":"1736302926005"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302926005"}]},"ts":"1736302926005"} 2025-01-08T02:22:06,007 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:22:06,008 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:22:06,008 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302926008"}]},"ts":"1736302926008"} 2025-01-08T02:22:06,010 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=ENABLING in hbase:meta 2025-01-08T02:22:06,013 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:22:06,014 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:22:06,014 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:22:06,014 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:22:06,014 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:22:06,014 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:22:06,014 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:22:06,014 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:22:06,015 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=1a942e02b6f510a0af53f37e83ed048b, ASSIGN}, {pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=b0d1552d5ae3c896125d7ba16ec8a867, ASSIGN}] 2025-01-08T02:22:06,016 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=b0d1552d5ae3c896125d7ba16ec8a867, ASSIGN 2025-01-08T02:22:06,016 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=1a942e02b6f510a0af53f37e83ed048b, ASSIGN 2025-01-08T02:22:06,017 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=100, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=1a942e02b6f510a0af53f37e83ed048b, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:22:06,017 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=101, ppid=99, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=b0d1552d5ae3c896125d7ba16ec8a867, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:22:06,065 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T02:22:06,168 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:22:06,168 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=100 updating hbase:meta row=1a942e02b6f510a0af53f37e83ed048b, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:06,168 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=101 updating hbase:meta row=b0d1552d5ae3c896125d7ba16ec8a867, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:06,172 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=102, ppid=100, state=RUNNABLE; OpenRegionProcedure 1a942e02b6f510a0af53f37e83ed048b, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:06,173 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=103, ppid=101, state=RUNNABLE; OpenRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:06,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T02:22:06,328 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:06,330 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:06,334 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,334 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7285): Opening region: {ENCODED => b0d1552d5ae3c896125d7ba16ec8a867, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:22:06,335 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. service=AccessControlService 2025-01-08T02:22:06,335 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:06,335 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,335 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:06,335 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7327): checking encryption for b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,336 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(7330): checking classloading for b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,341 INFO [StoreOpener-b0d1552d5ae3c896125d7ba16ec8a867-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,342 INFO [StoreOpener-b0d1552d5ae3c896125d7ba16ec8a867-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region b0d1552d5ae3c896125d7ba16ec8a867 columnFamilyName cf 2025-01-08T02:22:06,344 DEBUG [StoreOpener-b0d1552d5ae3c896125d7ba16ec8a867-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:06,344 INFO [StoreOpener-b0d1552d5ae3c896125d7ba16ec8a867-1 {}] regionserver.HStore(327): Store=b0d1552d5ae3c896125d7ba16ec8a867/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:06,346 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,346 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,347 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,347 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7285): Opening region: {ENCODED => 1a942e02b6f510a0af53f37e83ed048b, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:22:06,347 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. service=AccessControlService 2025-01-08T02:22:06,347 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:06,347 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,348 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:06,348 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7327): checking encryption for 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,348 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(7330): checking classloading for 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,349 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1085): writing seq id for b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,357 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:06,358 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1102): Opened b0d1552d5ae3c896125d7ba16ec8a867; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=62980011, jitterRate=-0.06152470409870148}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:06,359 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegion(1001): Region open journal for b0d1552d5ae3c896125d7ba16ec8a867: 2025-01-08T02:22:06,360 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867., pid=103, masterSystemTime=1736302926330 2025-01-08T02:22:06,362 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,362 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=103}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,362 INFO [StoreOpener-1a942e02b6f510a0af53f37e83ed048b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,363 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=101 updating hbase:meta row=b0d1552d5ae3c896125d7ba16ec8a867, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:06,364 INFO [StoreOpener-1a942e02b6f510a0af53f37e83ed048b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 1a942e02b6f510a0af53f37e83ed048b columnFamilyName cf 2025-01-08T02:22:06,365 DEBUG [StoreOpener-1a942e02b6f510a0af53f37e83ed048b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:06,365 INFO [StoreOpener-1a942e02b6f510a0af53f37e83ed048b-1 {}] regionserver.HStore(327): Store=1a942e02b6f510a0af53f37e83ed048b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:06,367 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,367 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,367 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=103, resume processing ppid=101 2025-01-08T02:22:06,367 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=103, ppid=101, state=SUCCESS; OpenRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867, server=c4e5d52b9759,39911,1736302808913 in 192 msec 2025-01-08T02:22:06,369 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=101, ppid=99, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=b0d1552d5ae3c896125d7ba16ec8a867, ASSIGN in 353 msec 2025-01-08T02:22:06,370 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1085): writing seq id for 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,372 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:06,373 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1102): Opened 1a942e02b6f510a0af53f37e83ed048b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63144106, jitterRate=-0.059079498052597046}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:06,373 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegion(1001): Region open journal for 1a942e02b6f510a0af53f37e83ed048b: 2025-01-08T02:22:06,374 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b., pid=102, masterSystemTime=1736302926328 2025-01-08T02:22:06,376 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,376 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=102}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,376 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=100 updating hbase:meta row=1a942e02b6f510a0af53f37e83ed048b, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:06,380 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=102, resume processing ppid=100 2025-01-08T02:22:06,380 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=102, ppid=100, state=SUCCESS; OpenRegionProcedure 1a942e02b6f510a0af53f37e83ed048b, server=c4e5d52b9759,37031,1736302808764 in 206 msec 2025-01-08T02:22:06,382 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=100, resume processing ppid=99 2025-01-08T02:22:06,382 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=100, ppid=99, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=1a942e02b6f510a0af53f37e83ed048b, ASSIGN in 366 msec 2025-01-08T02:22:06,383 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:22:06,383 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302926383"}]},"ts":"1736302926383"} 2025-01-08T02:22:06,385 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=ENABLED in hbase:meta 2025-01-08T02:22:06,388 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=99, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:22:06,389 DEBUG [PEWorker-4 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithMergeRegion jenkins: RWXCA 2025-01-08T02:22:06,389 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:22:06,392 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T02:22:06,394 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:06,394 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:06,394 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:06,395 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:06,397 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:06,397 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:06,397 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:06,398 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:06,399 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=99, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 436 msec 2025-01-08T02:22:06,568 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=99 2025-01-08T02:22:06,568 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 99 completed 2025-01-08T02:22:06,571 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:06,572 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,572 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:06,602 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T02:22:06,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302926602 (current time:1736302926602). 2025-01-08T02:22:06,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:06,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion VERSION not specified, setting to 2 2025-01-08T02:22:06,602 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:06,603 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x412c8575 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@63fce24a 2025-01-08T02:22:06,613 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@23b50b0e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:06,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:06,616 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37730, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:06,617 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x412c8575 to 127.0.0.1:61593 2025-01-08T02:22:06,617 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:06,619 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x13f104cc to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@179b4a4a 2025-01-08T02:22:06,627 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@1c7ddecc, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:06,629 DEBUG [hconnection-0x33c314c5-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:06,630 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37736, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:06,632 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:06,633 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54528, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:06,634 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x13f104cc to 127.0.0.1:61593 2025-01-08T02:22:06,634 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:06,634 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T02:22:06,635 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:06,635 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=104, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T02:22:06,636 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 104 2025-01-08T02:22:06,636 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:06,637 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T02:22:06,637 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:06,639 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:06,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742127_1303 (size=215) 2025-01-08T02:22:06,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742127_1303 (size=215) 2025-01-08T02:22:06,651 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742127_1303 (size=215) 2025-01-08T02:22:06,652 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:06,652 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b}, {pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867}] 2025-01-08T02:22:06,653 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,653 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,738 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T02:22:06,804 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:06,804 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:06,805 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=105 2025-01-08T02:22:06,805 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=106 2025-01-08T02:22:06,805 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,805 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.HRegion(2538): Flush status journal for 1a942e02b6f510a0af53f37e83ed048b: 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.HRegion(2538): Flush status journal for b0d1552d5ae3c896125d7ba16ec8a867: 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. for emptySnaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. for emptySnaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:06,806 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:06,828 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742128_1304 (size=86) 2025-01-08T02:22:06,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742129_1305 (size=86) 2025-01-08T02:22:06,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742128_1304 (size=86) 2025-01-08T02:22:06,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742128_1304 (size=86) 2025-01-08T02:22:06,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:06,830 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=105}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=105 2025-01-08T02:22:06,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742129_1305 (size=86) 2025-01-08T02:22:06,831 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742129_1305 (size=86) 2025-01-08T02:22:06,831 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=105 2025-01-08T02:22:06,831 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion on region 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,831 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:06,831 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=105, ppid=104, state=RUNNABLE; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:06,831 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=106}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=106 2025-01-08T02:22:06,832 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=106 2025-01-08T02:22:06,832 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithMergeRegion on region b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,832 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=106, ppid=104, state=RUNNABLE; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:06,834 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=105, ppid=104, state=SUCCESS; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b in 180 msec 2025-01-08T02:22:06,835 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=106, resume processing ppid=104 2025-01-08T02:22:06,835 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=106, ppid=104, state=SUCCESS; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867 in 181 msec 2025-01-08T02:22:06,835 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:06,836 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:06,837 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:06,837 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:06,838 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:06,838 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:22:06,856 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742130_1306 (size=78) 2025-01-08T02:22:06,856 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742130_1306 (size=78) 2025-01-08T02:22:06,857 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742130_1306 (size=78) 2025-01-08T02:22:06,859 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:06,859 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:06,860 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:06,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742131_1307 (size=713) 2025-01-08T02:22:06,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742131_1307 (size=713) 2025-01-08T02:22:06,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742131_1307 (size=713) 2025-01-08T02:22:06,939 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T02:22:06,941 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:06,964 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:06,969 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithMergeRegion to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:06,971 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=104, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:06,971 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 104 2025-01-08T02:22:06,972 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=104, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=104, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } in 336 msec 2025-01-08T02:22:07,241 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=104 2025-01-08T02:22:07,241 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 104 completed 2025-01-08T02:22:07,243 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:07,244 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:07,247 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,247 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:07,248 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:07,259 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T02:22:07,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302927259 (current time:1736302927259). 2025-01-08T02:22:07,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:07,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithMergeRegion VERSION not specified, setting to 2 2025-01-08T02:22:07,259 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:07,261 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x17724e97 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@3cbc0580 2025-01-08T02:22:07,267 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@69ee9c1b, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:07,268 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:07,269 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37744, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:07,270 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x17724e97 to 127.0.0.1:61593 2025-01-08T02:22:07,270 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:07,271 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x52628db8 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@e2e0271 2025-01-08T02:22:07,278 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@13bd49eb, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:07,284 DEBUG [hconnection-0x4f39cc66-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:07,285 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37758, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:07,287 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:07,288 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54532, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:07,288 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x52628db8 to 127.0.0.1:61593 2025-01-08T02:22:07,288 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:07,289 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion], kv [jenkins: RWXCA] 2025-01-08T02:22:07,290 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:07,290 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=107, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } 2025-01-08T02:22:07,291 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 107 2025-01-08T02:22:07,291 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:07,291 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T02:22:07,292 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:07,294 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:07,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742132_1308 (size=210) 2025-01-08T02:22:07,305 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742132_1308 (size=210) 2025-01-08T02:22:07,305 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742132_1308 (size=210) 2025-01-08T02:22:07,306 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:07,306 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b}, {pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867}] 2025-01-08T02:22:07,307 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:07,307 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:07,393 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T02:22:07,458 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:07,458 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:07,459 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=108 2025-01-08T02:22:07,459 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=109 2025-01-08T02:22:07,459 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:07,459 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:07,459 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(2837): Flushing 1a942e02b6f510a0af53f37e83ed048b 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T02:22:07,459 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(2837): Flushing b0d1552d5ae3c896125d7ba16ec8a867 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T02:22:07,479 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010862f0679b60e7487298d845a3ac712400_b0d1552d5ae3c896125d7ba16ec8a867 is 71, key is 15dc5f6a02d0aff44d359fe928a06395/cf:q/1736302927244/Put/seqid=0 2025-01-08T02:22:07,483 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108b6ba1d2b20b94ec98c04e318baf29791_1a942e02b6f510a0af53f37e83ed048b is 71, key is 007424152539e027288cd241058fdda2/cf:q/1736302927243/Put/seqid=0 2025-01-08T02:22:07,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742133_1309 (size=8101) 2025-01-08T02:22:07,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742133_1309 (size=8101) 2025-01-08T02:22:07,485 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742133_1309 (size=8101) 2025-01-08T02:22:07,486 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:07,488 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742134_1310 (size=5172) 2025-01-08T02:22:07,489 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742134_1310 (size=5172) 2025-01-08T02:22:07,489 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742134_1310 (size=5172) 2025-01-08T02:22:07,490 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:07,492 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010862f0679b60e7487298d845a3ac712400_b0d1552d5ae3c896125d7ba16ec8a867 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010862f0679b60e7487298d845a3ac712400_b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:07,493 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/.tmp/cf/4e36297119b449e699797745b67d736b, store: [table=testtb-testExportFileSystemStateWithMergeRegion family=cf region=b0d1552d5ae3c896125d7ba16ec8a867] 2025-01-08T02:22:07,494 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/.tmp/cf/4e36297119b449e699797745b67d736b is 224, key is 1feeee114ed9440540e32db04601d0e65/cf:q/1736302927244/Put/seqid=0 2025-01-08T02:22:07,494 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108b6ba1d2b20b94ec98c04e318baf29791_1a942e02b6f510a0af53f37e83ed048b to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108b6ba1d2b20b94ec98c04e318baf29791_1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:07,495 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/.tmp/cf/74d76542c55e41ae8775c468ab4c4f79, store: [table=testtb-testExportFileSystemStateWithMergeRegion family=cf region=1a942e02b6f510a0af53f37e83ed048b] 2025-01-08T02:22:07,496 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/.tmp/cf/74d76542c55e41ae8775c468ab4c4f79 is 224, key is 0cd25e714dee8152fba4b3da16f108655/cf:q/1736302927243/Put/seqid=0 2025-01-08T02:22:07,499 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742135_1311 (size=15497) 2025-01-08T02:22:07,499 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742135_1311 (size=15497) 2025-01-08T02:22:07,500 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742135_1311 (size=15497) 2025-01-08T02:22:07,500 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/.tmp/cf/4e36297119b449e699797745b67d736b 2025-01-08T02:22:07,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742136_1312 (size=6198) 2025-01-08T02:22:07,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742136_1312 (size=6198) 2025-01-08T02:22:07,503 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742136_1312 (size=6198) 2025-01-08T02:22:07,503 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/.tmp/cf/74d76542c55e41ae8775c468ab4c4f79 2025-01-08T02:22:07,506 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/.tmp/cf/4e36297119b449e699797745b67d736b as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/cf/4e36297119b449e699797745b67d736b 2025-01-08T02:22:07,508 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/.tmp/cf/74d76542c55e41ae8775c468ab4c4f79 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/cf/74d76542c55e41ae8775c468ab4c4f79 2025-01-08T02:22:07,511 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/cf/4e36297119b449e699797745b67d736b, entries=46, sequenceid=6, filesize=15.1 K 2025-01-08T02:22:07,512 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for b0d1552d5ae3c896125d7ba16ec8a867 in 53ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:07,512 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithMergeRegion' 2025-01-08T02:22:07,512 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/cf/74d76542c55e41ae8775c468ab4c4f79, entries=4, sequenceid=6, filesize=6.1 K 2025-01-08T02:22:07,512 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.HRegion(2538): Flush status journal for b0d1552d5ae3c896125d7ba16ec8a867: 2025-01-08T02:22:07,512 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. for snaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T02:22:07,512 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/cf/4e36297119b449e699797745b67d736b] hfiles 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/cf/4e36297119b449e699797745b67d736b for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,513 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 1a942e02b6f510a0af53f37e83ed048b in 54ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.HRegion(2538): Flush status journal for 1a942e02b6f510a0af53f37e83ed048b: 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. for snaptb0-testExportFileSystemStateWithMergeRegion completed. 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/cf/74d76542c55e41ae8775c468ab4c4f79] hfiles 2025-01-08T02:22:07,513 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/cf/74d76542c55e41ae8775c468ab4c4f79 for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,522 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742137_1313 (size=125) 2025-01-08T02:22:07,522 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742137_1313 (size=125) 2025-01-08T02:22:07,523 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742137_1313 (size=125) 2025-01-08T02:22:07,523 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:07,523 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=109}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=109 2025-01-08T02:22:07,524 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742138_1314 (size=125) 2025-01-08T02:22:07,524 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742138_1314 (size=125) 2025-01-08T02:22:07,524 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=109 2025-01-08T02:22:07,524 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion on region b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:07,524 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=109, ppid=107, state=RUNNABLE; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:07,524 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742138_1314 (size=125) 2025-01-08T02:22:07,525 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:07,525 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=108}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=108 2025-01-08T02:22:07,525 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=108 2025-01-08T02:22:07,525 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion on region 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:07,526 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=108, ppid=107, state=RUNNABLE; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:07,526 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=109, ppid=107, state=SUCCESS; SnapshotRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867 in 219 msec 2025-01-08T02:22:07,528 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=108, resume processing ppid=107 2025-01-08T02:22:07,528 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:07,528 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=108, ppid=107, state=SUCCESS; SnapshotRegionProcedure 1a942e02b6f510a0af53f37e83ed048b in 220 msec 2025-01-08T02:22:07,528 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:07,529 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:07,529 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:07,529 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:07,530 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010862f0679b60e7487298d845a3ac712400_b0d1552d5ae3c896125d7ba16ec8a867, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108b6ba1d2b20b94ec98c04e318baf29791_1a942e02b6f510a0af53f37e83ed048b] hfiles 2025-01-08T02:22:07,530 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010862f0679b60e7487298d845a3ac712400_b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:07,530 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108b6ba1d2b20b94ec98c04e318baf29791_1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:07,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742139_1315 (size=309) 2025-01-08T02:22:07,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742139_1315 (size=309) 2025-01-08T02:22:07,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742139_1315 (size=309) 2025-01-08T02:22:07,537 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:07,537 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,538 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,545 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742140_1316 (size=1023) 2025-01-08T02:22:07,546 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742140_1316 (size=1023) 2025-01-08T02:22:07,546 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742140_1316 (size=1023) 2025-01-08T02:22:07,548 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:07,554 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:07,554 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:07,556 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=107, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:07,556 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 }, snapshot procedure id = 107 2025-01-08T02:22:07,557 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=107, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=107, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion table=testtb-testExportFileSystemStateWithMergeRegion type=FLUSH ttl=0 } in 266 msec 2025-01-08T02:22:07,594 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=107 2025-01-08T02:22:07,594 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 107 completed 2025-01-08T02:22:07,614 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T02:22:07,617 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37760, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T02:22:07,617 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T02:22:07,618 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T02:22:07,619 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54534, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T02:22:07,620 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T02:22:07,620 DEBUG [Time-limited test {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=AdminService, sasl=false 2025-01-08T02:22:07,621 INFO [RS-EventLoopGroup-5-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:43726, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=AdminService 2025-01-08T02:22:07,622 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.CompactSplit(323): Interrupting running compactions because user switched off compactions 2025-01-08T02:22:07,623 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemStateWithMergeRegion-1', {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:22:07,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=110, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:07,625 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:22:07,625 DEBUG [PEWorker-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:07,625 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithMergeRegion-1" procId is: 110 2025-01-08T02:22:07,626 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:22:07,626 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T02:22:07,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742141_1317 (size=399) 2025-01-08T02:22:07,636 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742141_1317 (size=399) 2025-01-08T02:22:07,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742141_1317 (size=399) 2025-01-08T02:22:07,638 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => c5e2192bbba02a81bf3003d522b64c9b, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b.', STARTKEY => '', ENDKEY => '2'}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion-1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:07,639 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 6b078cfa4af4c3d42f8fa42264ae001f, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f.', STARTKEY => '2', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithMergeRegion-1', {TABLE_ATTRIBUTES => {METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:07,646 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742142_1318 (size=85) 2025-01-08T02:22:07,646 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742142_1318 (size=85) 2025-01-08T02:22:07,647 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742142_1318 (size=85) 2025-01-08T02:22:07,647 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:07,647 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1681): Closing c5e2192bbba02a81bf3003d522b64c9b, disabling compactions & flushes 2025-01-08T02:22:07,647 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:07,647 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:07,647 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. after waiting 0 ms 2025-01-08T02:22:07,647 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:07,647 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:07,647 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-0 {}] regionserver.HRegion(1635): Region close journal for c5e2192bbba02a81bf3003d522b64c9b: 2025-01-08T02:22:07,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742143_1319 (size=85) 2025-01-08T02:22:07,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742143_1319 (size=85) 2025-01-08T02:22:07,650 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742143_1319 (size=85) 2025-01-08T02:22:07,651 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:07,651 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1681): Closing 6b078cfa4af4c3d42f8fa42264ae001f, disabling compactions & flushes 2025-01-08T02:22:07,651 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:07,651 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:07,651 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. after waiting 0 ms 2025-01-08T02:22:07,651 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:07,651 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:07,651 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithMergeRegion-1-pool-1 {}] regionserver.HRegion(1635): Region close journal for 6b078cfa4af4c3d42f8fa42264ae001f: 2025-01-08T02:22:07,652 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:22:07,653 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b.","families":{"info":[{"qualifier":"regioninfo","vlen":84,"tag":[],"timestamp":"1736302927652"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302927652"}]},"ts":"1736302927652"} 2025-01-08T02:22:07,653 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f.","families":{"info":[{"qualifier":"regioninfo","vlen":84,"tag":[],"timestamp":"1736302927652"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302927652"}]},"ts":"1736302927652"} 2025-01-08T02:22:07,655 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:22:07,656 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:22:07,656 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302927656"}]},"ts":"1736302927656"} 2025-01-08T02:22:07,657 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=ENABLING in hbase:meta 2025-01-08T02:22:07,661 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:22:07,662 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:22:07,662 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:22:07,662 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:22:07,662 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:22:07,662 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:22:07,662 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:22:07,662 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:22:07,662 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=c5e2192bbba02a81bf3003d522b64c9b, ASSIGN}, {pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=6b078cfa4af4c3d42f8fa42264ae001f, ASSIGN}] 2025-01-08T02:22:07,663 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=6b078cfa4af4c3d42f8fa42264ae001f, ASSIGN 2025-01-08T02:22:07,663 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=c5e2192bbba02a81bf3003d522b64c9b, ASSIGN 2025-01-08T02:22:07,664 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=112, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=6b078cfa4af4c3d42f8fa42264ae001f, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:22:07,664 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=111, ppid=110, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=c5e2192bbba02a81bf3003d522b64c9b, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:22:07,727 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T02:22:07,814 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:22:07,815 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=112 updating hbase:meta row=6b078cfa4af4c3d42f8fa42264ae001f, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:07,815 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=111 updating hbase:meta row=c5e2192bbba02a81bf3003d522b64c9b, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:07,816 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=113, ppid=112, state=RUNNABLE; OpenRegionProcedure 6b078cfa4af4c3d42f8fa42264ae001f, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:07,817 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=114, ppid=111, state=RUNNABLE; OpenRegionProcedure c5e2192bbba02a81bf3003d522b64c9b, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:07,928 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T02:22:07,968 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:07,968 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:07,971 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:07,971 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:07,971 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7285): Opening region: {ENCODED => 6b078cfa4af4c3d42f8fa42264ae001f, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f.', STARTKEY => '2', ENDKEY => ''} 2025-01-08T02:22:07,971 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7285): Opening region: {ENCODED => c5e2192bbba02a81bf3003d522b64c9b, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b.', STARTKEY => '', ENDKEY => '2'} 2025-01-08T02:22:07,971 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. service=AccessControlService 2025-01-08T02:22:07,971 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. service=AccessControlService 2025-01-08T02:22:07,971 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:07,971 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:07,971 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:07,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:07,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7327): checking encryption for c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:07,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:07,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(7330): checking classloading for c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:07,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:07,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7327): checking encryption for 6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:07,972 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(7330): checking classloading for 6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:07,973 INFO [StoreOpener-c5e2192bbba02a81bf3003d522b64c9b-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:07,973 INFO [StoreOpener-6b078cfa4af4c3d42f8fa42264ae001f-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:07,975 INFO [StoreOpener-6b078cfa4af4c3d42f8fa42264ae001f-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 6b078cfa4af4c3d42f8fa42264ae001f columnFamilyName cf 2025-01-08T02:22:07,975 INFO [StoreOpener-c5e2192bbba02a81bf3003d522b64c9b-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region c5e2192bbba02a81bf3003d522b64c9b columnFamilyName cf 2025-01-08T02:22:07,975 DEBUG [StoreOpener-6b078cfa4af4c3d42f8fa42264ae001f-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:07,975 DEBUG [StoreOpener-c5e2192bbba02a81bf3003d522b64c9b-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:07,976 INFO [StoreOpener-6b078cfa4af4c3d42f8fa42264ae001f-1 {}] regionserver.HStore(327): Store=6b078cfa4af4c3d42f8fa42264ae001f/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:07,976 INFO [StoreOpener-c5e2192bbba02a81bf3003d522b64c9b-1 {}] regionserver.HStore(327): Store=c5e2192bbba02a81bf3003d522b64c9b/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:07,976 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:07,976 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:07,977 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:07,977 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:07,978 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1085): writing seq id for 6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:07,979 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1085): writing seq id for c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:07,980 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:07,981 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1102): Opened 6b078cfa4af4c3d42f8fa42264ae001f; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=69015214, jitterRate=0.02840682864189148}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:07,981 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:07,981 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1102): Opened c5e2192bbba02a81bf3003d522b64c9b; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72515661, jitterRate=0.08056755363941193}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:07,982 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegion(1001): Region open journal for 6b078cfa4af4c3d42f8fa42264ae001f: 2025-01-08T02:22:07,982 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegion(1001): Region open journal for c5e2192bbba02a81bf3003d522b64c9b: 2025-01-08T02:22:07,982 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b., pid=114, masterSystemTime=1736302927968 2025-01-08T02:22:07,983 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f., pid=113, masterSystemTime=1736302927968 2025-01-08T02:22:07,984 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:07,984 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=114}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:07,984 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=111 updating hbase:meta row=c5e2192bbba02a81bf3003d522b64c9b, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:07,985 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:07,985 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=113}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:07,985 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=112 updating hbase:meta row=6b078cfa4af4c3d42f8fa42264ae001f, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:07,987 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=114, resume processing ppid=111 2025-01-08T02:22:07,987 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=114, ppid=111, state=SUCCESS; OpenRegionProcedure c5e2192bbba02a81bf3003d522b64c9b, server=c4e5d52b9759,39911,1736302808913 in 169 msec 2025-01-08T02:22:07,988 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=113, resume processing ppid=112 2025-01-08T02:22:07,988 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=113, ppid=112, state=SUCCESS; OpenRegionProcedure 6b078cfa4af4c3d42f8fa42264ae001f, server=c4e5d52b9759,37031,1736302808764 in 170 msec 2025-01-08T02:22:07,989 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=111, ppid=110, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=c5e2192bbba02a81bf3003d522b64c9b, ASSIGN in 325 msec 2025-01-08T02:22:07,989 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=112, resume processing ppid=110 2025-01-08T02:22:07,990 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=112, ppid=110, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=6b078cfa4af4c3d42f8fa42264ae001f, ASSIGN in 326 msec 2025-01-08T02:22:07,990 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:22:07,990 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302927990"}]},"ts":"1736302927990"} 2025-01-08T02:22:07,991 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=ENABLED in hbase:meta 2025-01-08T02:22:07,994 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=110, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:22:07,994 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithMergeRegion-1 jenkins: RWXCA 2025-01-08T02:22:07,996 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion-1], kv [jenkins: RWXCA] 2025-01-08T02:22:07,997 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:07,998 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:07,998 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:07,998 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,000 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF\x0AW\x0A\x07jenkins\x12L\x08\x03"H\x0A<\x0A\x07default\x121testtb-testExportFileSystemStateWithMergeRegion-1 \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:08,001 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=110, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 376 msec 2025-01-08T02:22:08,229 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=110 2025-01-08T02:22:08,229 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 110 completed 2025-01-08T02:22:08,246 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$2(2234): Client=jenkins//172.17.0.3 merge regions [c5e2192bbba02a81bf3003d522b64c9b, 6b078cfa4af4c3d42f8fa42264ae001f] 2025-01-08T02:22:08,251 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[c5e2192bbba02a81bf3003d522b64c9b, 6b078cfa4af4c3d42f8fa42264ae001f], force=true 2025-01-08T02:22:08,252 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[c5e2192bbba02a81bf3003d522b64c9b, 6b078cfa4af4c3d42f8fa42264ae001f], force=true 2025-01-08T02:22:08,252 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[c5e2192bbba02a81bf3003d522b64c9b, 6b078cfa4af4c3d42f8fa42264ae001f], force=true 2025-01-08T02:22:08,252 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=115, state=RUNNABLE:MERGE_TABLE_REGIONS_PREPARE; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[c5e2192bbba02a81bf3003d522b64c9b, 6b078cfa4af4c3d42f8fa42264ae001f], force=true 2025-01-08T02:22:08,253 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T02:22:08,265 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=116, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=c5e2192bbba02a81bf3003d522b64c9b, UNASSIGN}, {pid=117, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=6b078cfa4af4c3d42f8fa42264ae001f, UNASSIGN}] 2025-01-08T02:22:08,266 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=116, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=c5e2192bbba02a81bf3003d522b64c9b, UNASSIGN 2025-01-08T02:22:08,266 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=117, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=6b078cfa4af4c3d42f8fa42264ae001f, UNASSIGN 2025-01-08T02:22:08,266 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=116 updating hbase:meta row=c5e2192bbba02a81bf3003d522b64c9b, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:08,266 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=117 updating hbase:meta row=6b078cfa4af4c3d42f8fa42264ae001f, regionState=CLOSING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:08,268 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:08,268 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=118, ppid=116, state=RUNNABLE; CloseRegionProcedure c5e2192bbba02a81bf3003d522b64c9b, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:08,268 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: true: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:08,268 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=119, ppid=117, state=RUNNABLE; CloseRegionProcedure 6b078cfa4af4c3d42f8fa42264ae001f, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:08,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T02:22:08,420 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:08,420 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:08,420 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(124): Close c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:08,420 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(124): Close 6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:08,420 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2025-01-08T02:22:08,420 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(138): Unassign region: split region: true: evictCache: true 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1681): Closing c5e2192bbba02a81bf3003d522b64c9b, disabling compactions & flushes 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1681): Closing 6b078cfa4af4c3d42f8fa42264ae001f, disabling compactions & flushes 2025-01-08T02:22:08,421 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:08,421 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. after waiting 0 ms 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. after waiting 0 ms 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:08,421 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:08,421 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(2837): Flushing 6b078cfa4af4c3d42f8fa42264ae001f 1/1 column families, dataSize=24 B heapSize=352 B 2025-01-08T02:22:08,421 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(2837): Flushing c5e2192bbba02a81bf3003d522b64c9b 1/1 column families, dataSize=24 B heapSize=352 B 2025-01-08T02:22:08,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:08,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion Metrics about Tables on a single HBase RegionServer 2025-01-08T02:22:08,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:08,432 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 Metrics about Tables on a single HBase RegionServer 2025-01-08T02:22:08,433 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testConsecutiveExports 2025-01-08T02:22:08,437 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/.tmp/cf/3ff581bafb0b4a758e2e85f3e593c8b8 is 28, key is 2/cf:/1736302928235/Put/seqid=0 2025-01-08T02:22:08,438 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/.tmp/cf/5330a89c83ef4074926d1e1630ba855c is 28, key is 1/cf:/1736302928232/Put/seqid=0 2025-01-08T02:22:08,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742144_1320 (size=4945) 2025-01-08T02:22:08,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742145_1321 (size=4945) 2025-01-08T02:22:08,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742144_1320 (size=4945) 2025-01-08T02:22:08,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742145_1321 (size=4945) 2025-01-08T02:22:08,451 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742145_1321 (size=4945) 2025-01-08T02:22:08,452 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24 B at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/.tmp/cf/5330a89c83ef4074926d1e1630ba855c 2025-01-08T02:22:08,452 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742144_1320 (size=4945) 2025-01-08T02:22:08,452 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=24 B at sequenceid=5 (bloomFilter=true), to=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/.tmp/cf/3ff581bafb0b4a758e2e85f3e593c8b8 2025-01-08T02:22:08,457 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/.tmp/cf/5330a89c83ef4074926d1e1630ba855c as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/cf/5330a89c83ef4074926d1e1630ba855c 2025-01-08T02:22:08,458 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/.tmp/cf/3ff581bafb0b4a758e2e85f3e593c8b8 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/cf/3ff581bafb0b4a758e2e85f3e593c8b8 2025-01-08T02:22:08,462 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/cf/5330a89c83ef4074926d1e1630ba855c, entries=1, sequenceid=5, filesize=4.8 K 2025-01-08T02:22:08,462 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/cf/3ff581bafb0b4a758e2e85f3e593c8b8, entries=1, sequenceid=5, filesize=4.8 K 2025-01-08T02:22:08,463 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(3040): Finished flush of dataSize ~24 B/24, heapSize ~336 B/336, currentSize=0 B/0 for c5e2192bbba02a81bf3003d522b64c9b in 42ms, sequenceid=5, compaction requested=false 2025-01-08T02:22:08,463 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithMergeRegion-1' 2025-01-08T02:22:08,463 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(3040): Finished flush of dataSize ~24 B/24, heapSize ~336 B/336, currentSize=0 B/0 for 6b078cfa4af4c3d42f8fa42264ae001f in 42ms, sequenceid=5, compaction requested=false 2025-01-08T02:22:08,463 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithMergeRegion-1' 2025-01-08T02:22:08,473 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T02:22:08,473 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T02:22:08,473 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:08,473 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:08,473 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f. 2025-01-08T02:22:08,474 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b. 2025-01-08T02:22:08,474 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] regionserver.HRegion(1635): Region close journal for 6b078cfa4af4c3d42f8fa42264ae001f: 2025-01-08T02:22:08,474 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] regionserver.HRegion(1635): Region close journal for c5e2192bbba02a81bf3003d522b64c9b: 2025-01-08T02:22:08,475 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=118}] handler.UnassignRegionHandler(170): Closed c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:08,475 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=116 updating hbase:meta row=c5e2192bbba02a81bf3003d522b64c9b, regionState=CLOSED 2025-01-08T02:22:08,475 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=119}] handler.UnassignRegionHandler(170): Closed 6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:08,476 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=117 updating hbase:meta row=6b078cfa4af4c3d42f8fa42264ae001f, regionState=CLOSED 2025-01-08T02:22:08,478 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=118, resume processing ppid=116 2025-01-08T02:22:08,479 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=119, resume processing ppid=117 2025-01-08T02:22:08,479 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=118, ppid=116, state=SUCCESS; CloseRegionProcedure c5e2192bbba02a81bf3003d522b64c9b, server=c4e5d52b9759,39911,1736302808913 in 209 msec 2025-01-08T02:22:08,479 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=116, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=c5e2192bbba02a81bf3003d522b64c9b, UNASSIGN in 213 msec 2025-01-08T02:22:08,479 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=119, ppid=117, state=SUCCESS; CloseRegionProcedure 6b078cfa4af4c3d42f8fa42264ae001f, server=c4e5d52b9759,37031,1736302808764 in 209 msec 2025-01-08T02:22:08,480 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=117, resume processing ppid=115 2025-01-08T02:22:08,480 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=117, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=6b078cfa4af4c3d42f8fa42264ae001f, UNASSIGN in 214 msec 2025-01-08T02:22:08,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742146_1322 (size=84) 2025-01-08T02:22:08,491 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742146_1322 (size=84) 2025-01-08T02:22:08,492 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742146_1322 (size=84) 2025-01-08T02:22:08,493 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:08,505 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742147_1323 (size=20) 2025-01-08T02:22:08,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742147_1323 (size=20) 2025-01-08T02:22:08,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742147_1323 (size=20) 2025-01-08T02:22:08,507 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:08,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742148_1324 (size=21) 2025-01-08T02:22:08,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742148_1324 (size=21) 2025-01-08T02:22:08,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742148_1324 (size=21) 2025-01-08T02:22:08,518 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742149_1325 (size=84) 2025-01-08T02:22:08,518 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742149_1325 (size=84) 2025-01-08T02:22:08,518 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742149_1325 (size=84) 2025-01-08T02:22:08,519 DEBUG [PEWorker-3 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:08,528 DEBUG [PEWorker-3 {}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=-1 2025-01-08T02:22:08,529 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927623.c5e2192bbba02a81bf3003d522b64c9b.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:08,529 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,2,1736302927623.6b078cfa4af4c3d42f8fa42264ae001f.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:08,529 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":7,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598.","families":{"info":[{"qualifier":"regioninfo","vlen":83,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"merge0000","vlen":84,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"merge0001","vlen":84,"tag":[],"timestamp":"9223372036854775807"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"9223372036854775807"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:08,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T02:22:08,558 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a25e31c91636f9ea9e4f60f60a12d598, ASSIGN}] 2025-01-08T02:22:08,559 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a25e31c91636f9ea9e4f60f60a12d598, ASSIGN 2025-01-08T02:22:08,560 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=120, ppid=115, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a25e31c91636f9ea9e4f60f60a12d598, ASSIGN; state=MERGED, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:22:08,710 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 1 regions. 1 retained the pre-restart assignment. 2025-01-08T02:22:08,710 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=120 updating hbase:meta row=a25e31c91636f9ea9e4f60f60a12d598, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:08,711 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=121, ppid=120, state=RUNNABLE; OpenRegionProcedure a25e31c91636f9ea9e4f60f60a12d598, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:08,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T02:22:08,863 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:08,866 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:08,866 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7285): Opening region: {ENCODED => a25e31c91636f9ea9e4f60f60a12d598, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598.', STARTKEY => '', ENDKEY => ''} 2025-01-08T02:22:08,867 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. service=AccessControlService 2025-01-08T02:22:08,867 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:08,867 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithMergeRegion-1 a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:08,867 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:08,867 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7327): checking encryption for a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:08,867 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(7330): checking classloading for a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:08,869 INFO [StoreOpener-a25e31c91636f9ea9e4f60f60a12d598-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:08,870 INFO [StoreOpener-a25e31c91636f9ea9e4f60f60a12d598-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region a25e31c91636f9ea9e4f60f60a12d598 columnFamilyName cf 2025-01-08T02:22:08,870 DEBUG [StoreOpener-a25e31c91636f9ea9e4f60f60a12d598-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:08,890 DEBUG [StoreOpener-a25e31c91636f9ea9e4f60f60a12d598-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/3ff581bafb0b4a758e2e85f3e593c8b8.6b078cfa4af4c3d42f8fa42264ae001f->hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/cf/3ff581bafb0b4a758e2e85f3e593c8b8-top 2025-01-08T02:22:08,896 DEBUG [StoreOpener-a25e31c91636f9ea9e4f60f60a12d598-1 {}] regionserver.StoreEngine(277): loaded hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/5330a89c83ef4074926d1e1630ba855c.c5e2192bbba02a81bf3003d522b64c9b->hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/cf/5330a89c83ef4074926d1e1630ba855c-top 2025-01-08T02:22:08,897 INFO [StoreOpener-a25e31c91636f9ea9e4f60f60a12d598-1 {}] regionserver.HStore(327): Store=a25e31c91636f9ea9e4f60f60a12d598/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:08,897 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:08,899 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:08,901 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1085): writing seq id for a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:08,901 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1102): Opened a25e31c91636f9ea9e4f60f60a12d598; next sequenceid=9; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=64244430, jitterRate=-0.04268339276313782}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:08,902 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegion(1001): Region open journal for a25e31c91636f9ea9e4f60f60a12d598: 2025-01-08T02:22:08,903 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598., pid=121, masterSystemTime=1736302928863 2025-01-08T02:22:08,904 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.CompactSplit(342): Ignoring compaction request for testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598.,because compaction is disabled. 2025-01-08T02:22:08,905 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:08,905 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=121}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:08,905 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=120 updating hbase:meta row=a25e31c91636f9ea9e4f60f60a12d598, regionState=OPEN, openSeqNum=9, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:08,908 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=121, resume processing ppid=120 2025-01-08T02:22:08,908 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=121, ppid=120, state=SUCCESS; OpenRegionProcedure a25e31c91636f9ea9e4f60f60a12d598, server=c4e5d52b9759,39911,1736302808913 in 196 msec 2025-01-08T02:22:08,910 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=120, resume processing ppid=115 2025-01-08T02:22:08,910 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=120, ppid=115, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a25e31c91636f9ea9e4f60f60a12d598, ASSIGN in 350 msec 2025-01-08T02:22:08,911 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=115, state=SUCCESS; MergeTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, regions=[c5e2192bbba02a81bf3003d522b64c9b, 6b078cfa4af4c3d42f8fa42264ae001f], force=true in 662 msec 2025-01-08T02:22:09,357 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=115 2025-01-08T02:22:09,358 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: MERGE_REGIONS, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 115 completed 2025-01-08T02:22:09,358 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } 2025-01-08T02:22:09,358 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302929358 (current time:1736302929358). 2025-01-08T02:22:09,358 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:09,358 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithMergeRegion-1 VERSION not specified, setting to 2 2025-01-08T02:22:09,358 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:09,359 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x1a581062 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@123d8229 2025-01-08T02:22:09,368 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@45ba5da3, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:09,370 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:09,371 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37776, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:09,372 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x1a581062 to 127.0.0.1:61593 2025-01-08T02:22:09,372 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:09,373 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x0034d1cf to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@7e66272a 2025-01-08T02:22:09,385 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@69a62a00, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:09,387 DEBUG [hconnection-0x1b6e4486-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:09,388 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:37778, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:09,390 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:09,391 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:54538, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:09,392 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x0034d1cf to 127.0.0.1:61593 2025-01-08T02:22:09,392 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:09,392 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithMergeRegion-1], kv [jenkins: RWXCA] 2025-01-08T02:22:09,393 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:09,393 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=122, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } 2025-01-08T02:22:09,394 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 }, snapshot procedure id = 122 2025-01-08T02:22:09,394 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:09,394 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T02:22:09,395 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:09,401 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:09,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742150_1326 (size=216) 2025-01-08T02:22:09,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742150_1326 (size=216) 2025-01-08T02:22:09,416 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742150_1326 (size=216) 2025-01-08T02:22:09,417 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:09,418 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure a25e31c91636f9ea9e4f60f60a12d598}] 2025-01-08T02:22:09,420 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:09,495 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T02:22:09,574 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:09,574 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=123 2025-01-08T02:22:09,574 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:09,575 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.HRegion(2538): Flush status journal for a25e31c91636f9ea9e4f60f60a12d598: 2025-01-08T02:22:09,575 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. for snaptb0-testExportFileSystemStateWithMergeRegion-1 completed. 2025-01-08T02:22:09,575 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598.' region-info for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,575 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:09,575 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/3ff581bafb0b4a758e2e85f3e593c8b8.6b078cfa4af4c3d42f8fa42264ae001f->hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/cf/3ff581bafb0b4a758e2e85f3e593c8b8-top, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/5330a89c83ef4074926d1e1630ba855c.c5e2192bbba02a81bf3003d522b64c9b->hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/cf/5330a89c83ef4074926d1e1630ba855c-top] hfiles 2025-01-08T02:22:09,575 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(265): Adding reference for file (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/3ff581bafb0b4a758e2e85f3e593c8b8.6b078cfa4af4c3d42f8fa42264ae001f for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,576 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] snapshot.SnapshotManifest(265): Adding reference for file (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/5330a89c83ef4074926d1e1630ba855c.c5e2192bbba02a81bf3003d522b64c9b for snapshot=snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742151_1327 (size=269) 2025-01-08T02:22:09,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742151_1327 (size=269) 2025-01-08T02:22:09,593 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742151_1327 (size=269) 2025-01-08T02:22:09,594 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:09,594 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=123}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=123 2025-01-08T02:22:09,594 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=123 2025-01-08T02:22:09,594 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithMergeRegion-1 on region a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:09,595 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=123, ppid=122, state=RUNNABLE; SnapshotRegionProcedure a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:09,597 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=123, resume processing ppid=122 2025-01-08T02:22:09,597 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=123, ppid=122, state=SUCCESS; SnapshotRegionProcedure a25e31c91636f9ea9e4f60f60a12d598 in 177 msec 2025-01-08T02:22:09,597 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:09,598 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:09,598 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:09,598 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,599 DEBUG [PEWorker-3 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742152_1328 (size=670) 2025-01-08T02:22:09,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742152_1328 (size=670) 2025-01-08T02:22:09,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742152_1328 (size=670) 2025-01-08T02:22:09,609 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:09,614 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:09,614 DEBUG [PEWorker-3 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,615 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=122, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:09,615 DEBUG [PEWorker-3 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 }, snapshot procedure id = 122 2025-01-08T02:22:09,616 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=122, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=122, snapshot={ ss=snaptb0-testExportFileSystemStateWithMergeRegion-1 table=testtb-testExportFileSystemStateWithMergeRegion-1 type=FLUSH ttl=0 } in 222 msec 2025-01-08T02:22:09,697 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=122 2025-01-08T02:22:09,697 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 122 completed 2025-01-08T02:22:09,697 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697 2025-01-08T02:22:09,697 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:09,746 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:09,746 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,747 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:22:09,752 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:09,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742154_1330 (size=670) 2025-01-08T02:22:09,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742154_1330 (size=670) 2025-01-08T02:22:09,765 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742154_1330 (size=670) 2025-01-08T02:22:09,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742153_1329 (size=216) 2025-01-08T02:22:09,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742153_1329 (size=216) 2025-01-08T02:22:09,769 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742153_1329 (size=216) 2025-01-08T02:22:09,771 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:09,771 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:09,771 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:09,771 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:10,558 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0005_000001 (auth:SIMPLE) from 127.0.0.1:59302 2025-01-08T02:22:10,583 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0005/container_1736302816589_0005_01_000001/launch_container.sh] 2025-01-08T02:22:10,583 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0005/container_1736302816589_0005_01_000001/container_tokens] 2025-01-08T02:22:10,583 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0005/container_1736302816589_0005_01_000001/sysfs] 2025-01-08T02:22:11,172 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-15881901925343986668.jar 2025-01-08T02:22:11,172 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,173 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,240 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:22:11,262 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-16525728989584506275.jar 2025-01-08T02:22:11,263 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,263 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,263 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,264 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,264 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,264 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:11,265 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:22:11,265 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:22:11,265 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:22:11,266 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:22:11,266 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:22:11,266 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:22:11,267 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:22:11,267 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:22:11,267 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:22:11,268 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:22:11,268 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:22:11,268 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:22:11,269 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:11,269 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:11,269 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:11,270 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:11,270 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:11,270 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:11,271 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:11,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742155_1331 (size=127628) 2025-01-08T02:22:11,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742155_1331 (size=127628) 2025-01-08T02:22:11,364 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742155_1331 (size=127628) 2025-01-08T02:22:11,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T02:22:11,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T02:22:11,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742156_1332 (size=2172137) 2025-01-08T02:22:11,413 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742157_1333 (size=213228) 2025-01-08T02:22:11,413 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742157_1333 (size=213228) 2025-01-08T02:22:11,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742157_1333 (size=213228) 2025-01-08T02:22:11,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T02:22:11,442 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T02:22:11,442 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742158_1334 (size=1877034) 2025-01-08T02:22:11,454 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742159_1335 (size=533455) 2025-01-08T02:22:11,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742159_1335 (size=533455) 2025-01-08T02:22:11,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742159_1335 (size=533455) 2025-01-08T02:22:11,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742160_1336 (size=7280644) 2025-01-08T02:22:11,506 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742160_1336 (size=7280644) 2025-01-08T02:22:11,507 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742160_1336 (size=7280644) 2025-01-08T02:22:11,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742161_1337 (size=4188619) 2025-01-08T02:22:11,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742161_1337 (size=4188619) 2025-01-08T02:22:11,536 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742161_1337 (size=4188619) 2025-01-08T02:22:11,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742162_1338 (size=20406) 2025-01-08T02:22:11,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742162_1338 (size=20406) 2025-01-08T02:22:11,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742162_1338 (size=20406) 2025-01-08T02:22:11,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742163_1339 (size=75495) 2025-01-08T02:22:11,588 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742163_1339 (size=75495) 2025-01-08T02:22:11,588 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742163_1339 (size=75495) 2025-01-08T02:22:11,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742164_1340 (size=45609) 2025-01-08T02:22:11,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742164_1340 (size=45609) 2025-01-08T02:22:11,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742164_1340 (size=45609) 2025-01-08T02:22:11,618 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742165_1341 (size=110084) 2025-01-08T02:22:11,619 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742165_1341 (size=110084) 2025-01-08T02:22:11,622 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 1a942e02b6f510a0af53f37e83ed048b changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:22:11,623 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region b0d1552d5ae3c896125d7ba16ec8a867 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:22:11,624 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742165_1341 (size=110084) 2025-01-08T02:22:11,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742166_1342 (size=451756) 2025-01-08T02:22:11,641 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742166_1342 (size=451756) 2025-01-08T02:22:11,642 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742166_1342 (size=451756) 2025-01-08T02:22:11,669 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742167_1343 (size=1323991) 2025-01-08T02:22:11,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742167_1343 (size=1323991) 2025-01-08T02:22:11,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742167_1343 (size=1323991) 2025-01-08T02:22:11,693 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742168_1344 (size=23076) 2025-01-08T02:22:11,693 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742168_1344 (size=23076) 2025-01-08T02:22:11,693 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742168_1344 (size=23076) 2025-01-08T02:22:12,106 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742169_1345 (size=126803) 2025-01-08T02:22:12,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742169_1345 (size=126803) 2025-01-08T02:22:12,107 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742169_1345 (size=126803) 2025-01-08T02:22:12,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742170_1346 (size=322274) 2025-01-08T02:22:12,119 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742170_1346 (size=322274) 2025-01-08T02:22:12,120 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742170_1346 (size=322274) 2025-01-08T02:22:12,145 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742171_1347 (size=1832290) 2025-01-08T02:22:12,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742171_1347 (size=1832290) 2025-01-08T02:22:12,146 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742171_1347 (size=1832290) 2025-01-08T02:22:12,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742172_1348 (size=30081) 2025-01-08T02:22:12,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742172_1348 (size=30081) 2025-01-08T02:22:12,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742172_1348 (size=30081) 2025-01-08T02:22:12,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742173_1349 (size=53616) 2025-01-08T02:22:12,173 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742173_1349 (size=53616) 2025-01-08T02:22:12,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742173_1349 (size=53616) 2025-01-08T02:22:12,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742174_1350 (size=6350715) 2025-01-08T02:22:12,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742174_1350 (size=6350715) 2025-01-08T02:22:12,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742174_1350 (size=6350715) 2025-01-08T02:22:12,213 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742175_1351 (size=29229) 2025-01-08T02:22:12,214 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742175_1351 (size=29229) 2025-01-08T02:22:12,214 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742175_1351 (size=29229) 2025-01-08T02:22:12,226 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742176_1352 (size=169089) 2025-01-08T02:22:12,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742176_1352 (size=169089) 2025-01-08T02:22:12,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742176_1352 (size=169089) 2025-01-08T02:22:12,262 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742177_1353 (size=5175431) 2025-01-08T02:22:12,263 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742177_1353 (size=5175431) 2025-01-08T02:22:12,263 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742177_1353 (size=5175431) 2025-01-08T02:22:12,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742178_1354 (size=136454) 2025-01-08T02:22:12,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742178_1354 (size=136454) 2025-01-08T02:22:12,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742178_1354 (size=136454) 2025-01-08T02:22:12,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742179_1355 (size=907469) 2025-01-08T02:22:12,290 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742179_1355 (size=907469) 2025-01-08T02:22:12,291 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742179_1355 (size=907469) 2025-01-08T02:22:12,310 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742180_1356 (size=3317408) 2025-01-08T02:22:12,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742180_1356 (size=3317408) 2025-01-08T02:22:12,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742180_1356 (size=3317408) 2025-01-08T02:22:12,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742181_1357 (size=503880) 2025-01-08T02:22:12,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742181_1357 (size=503880) 2025-01-08T02:22:12,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742181_1357 (size=503880) 2025-01-08T02:22:12,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T02:22:12,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T02:22:12,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742182_1358 (size=4695811) 2025-01-08T02:22:12,361 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:22:12,363 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemStateWithMergeRegion-1' hfile list 2025-01-08T02:22:12,364 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=9.7 K 2025-01-08T02:22:12,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742183_1359 (size=378) 2025-01-08T02:22:12,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742183_1359 (size=378) 2025-01-08T02:22:12,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742183_1359 (size=378) 2025-01-08T02:22:12,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742184_1360 (size=15) 2025-01-08T02:22:12,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742184_1360 (size=15) 2025-01-08T02:22:12,378 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742184_1360 (size=15) 2025-01-08T02:22:12,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742185_1361 (size=304940) 2025-01-08T02:22:12,391 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742185_1361 (size=304940) 2025-01-08T02:22:12,392 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742185_1361 (size=304940) 2025-01-08T02:22:12,409 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:22:12,409 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:22:12,534 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0006_000001 (auth:SIMPLE) from 127.0.0.1:59310 2025-01-08T02:22:17,792 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0006_000001 (auth:SIMPLE) from 127.0.0.1:46462 2025-01-08T02:22:18,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742186_1362 (size=350614) 2025-01-08T02:22:18,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742186_1362 (size=350614) 2025-01-08T02:22:18,038 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742186_1362 (size=350614) 2025-01-08T02:22:20,045 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0006_000001 (auth:SIMPLE) from 127.0.0.1:59812 2025-01-08T02:22:23,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742187_1363 (size=4945) 2025-01-08T02:22:23,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742187_1363 (size=4945) 2025-01-08T02:22:23,944 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742187_1363 (size=4945) 2025-01-08T02:22:24,009 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742188_1364 (size=4945) 2025-01-08T02:22:24,010 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742188_1364 (size=4945) 2025-01-08T02:22:24,011 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742188_1364 (size=4945) 2025-01-08T02:22:24,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742189_1365 (size=17474) 2025-01-08T02:22:24,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742189_1365 (size=17474) 2025-01-08T02:22:24,122 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742189_1365 (size=17474) 2025-01-08T02:22:24,162 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742190_1366 (size=482) 2025-01-08T02:22:24,164 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742190_1366 (size=482) 2025-01-08T02:22:24,175 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742190_1366 (size=482) 2025-01-08T02:22:24,212 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0006/container_1736302816589_0006_01_000002/launch_container.sh] 2025-01-08T02:22:24,212 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0006/container_1736302816589_0006_01_000002/container_tokens] 2025-01-08T02:22:24,212 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0006/container_1736302816589_0006_01_000002/sysfs] 2025-01-08T02:22:24,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742191_1367 (size=17474) 2025-01-08T02:22:24,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742191_1367 (size=17474) 2025-01-08T02:22:24,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742191_1367 (size=17474) 2025-01-08T02:22:24,272 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742192_1368 (size=350614) 2025-01-08T02:22:24,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742192_1368 (size=350614) 2025-01-08T02:22:24,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742192_1368 (size=350614) 2025-01-08T02:22:24,290 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0006_000001 (auth:SIMPLE) from 127.0.0.1:59822 2025-01-08T02:22:25,765 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:22:25,774 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:22:25,791 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:25,791 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:22:25,792 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:22:25,792 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:25,793 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/.snapshotinfo 2025-01-08T02:22:25,793 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/data.manifest 2025-01-08T02:22:25,793 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:25,794 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/.snapshotinfo 2025-01-08T02:22:25,794 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302929697/.hbase-snapshot/snaptb0-testExportFileSystemStateWithMergeRegion-1/data.manifest 2025-01-08T02:22:25,804 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:25,804 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:25,805 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=124, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:25,807 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T02:22:25,808 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302945807"}]},"ts":"1736302945807"} 2025-01-08T02:22:25,809 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=DISABLING in hbase:meta 2025-01-08T02:22:25,811 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithMergeRegion-1 to state=DISABLING 2025-01-08T02:22:25,812 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=125, ppid=124, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1}] 2025-01-08T02:22:25,813 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=126, ppid=125, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a25e31c91636f9ea9e4f60f60a12d598, UNASSIGN}] 2025-01-08T02:22:25,814 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=126, ppid=125, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a25e31c91636f9ea9e4f60f60a12d598, UNASSIGN 2025-01-08T02:22:25,815 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=126 updating hbase:meta row=a25e31c91636f9ea9e4f60f60a12d598, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:25,817 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:25,817 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=127, ppid=126, state=RUNNABLE; CloseRegionProcedure a25e31c91636f9ea9e4f60f60a12d598, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:25,909 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T02:22:25,968 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:25,969 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(124): Close a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:25,969 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:25,969 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1681): Closing a25e31c91636f9ea9e4f60f60a12d598, disabling compactions & flushes 2025-01-08T02:22:25,970 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:25,970 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:25,970 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. after waiting 0 ms 2025-01-08T02:22:25,970 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:25,975 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/recovered.edits/12.seqid, newMaxSeqId=12, maxSeqId=8 2025-01-08T02:22:25,975 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:25,975 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598. 2025-01-08T02:22:25,975 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] regionserver.HRegion(1635): Region close journal for a25e31c91636f9ea9e4f60f60a12d598: 2025-01-08T02:22:25,977 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=127}] handler.UnassignRegionHandler(170): Closed a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:25,977 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=126 updating hbase:meta row=a25e31c91636f9ea9e4f60f60a12d598, regionState=CLOSED 2025-01-08T02:22:25,980 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=127, resume processing ppid=126 2025-01-08T02:22:25,980 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=127, ppid=126, state=SUCCESS; CloseRegionProcedure a25e31c91636f9ea9e4f60f60a12d598, server=c4e5d52b9759,39911,1736302808913 in 162 msec 2025-01-08T02:22:25,982 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=126, resume processing ppid=125 2025-01-08T02:22:25,982 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=126, ppid=125, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1, region=a25e31c91636f9ea9e4f60f60a12d598, UNASSIGN in 167 msec 2025-01-08T02:22:25,984 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=125, resume processing ppid=124 2025-01-08T02:22:25,984 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=125, ppid=124, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 170 msec 2025-01-08T02:22:25,985 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302945985"}]},"ts":"1736302945985"} 2025-01-08T02:22:25,986 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion-1, state=DISABLED in hbase:meta 2025-01-08T02:22:25,989 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithMergeRegion-1 to state=DISABLED 2025-01-08T02:22:25,990 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=124, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 185 msec 2025-01-08T02:22:26,110 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=124 2025-01-08T02:22:26,110 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 124 completed 2025-01-08T02:22:26,111 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,111 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=128, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,112 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=128, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,112 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,113 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=128, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,114 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,116 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:26,116 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:26,116 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:26,117 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,117 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,117 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,118 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/recovered.edits] 2025-01-08T02:22:26,118 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/recovered.edits] 2025-01-08T02:22:26,118 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,118 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T02:22:26,118 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T02:22:26,118 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T02:22:26,118 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/recovered.edits] 2025-01-08T02:22:26,119 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion-1 with data PBUF 2025-01-08T02:22:26,119 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,120 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,121 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=128 2025-01-08T02:22:26,121 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:26,121 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:26,122 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:26,122 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF\x0AU\x0A\x07jenkins\x12J\x08\x03"F\x0A:\x0A\x07default\x12/testtb-testExportFileSystemStateWithMergeRegion \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:26,124 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/3ff581bafb0b4a758e2e85f3e593c8b8.6b078cfa4af4c3d42f8fa42264ae001f to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/3ff581bafb0b4a758e2e85f3e593c8b8.6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:26,124 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/cf/5330a89c83ef4074926d1e1630ba855c to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/cf/5330a89c83ef4074926d1e1630ba855c 2025-01-08T02:22:26,124 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/cf/3ff581bafb0b4a758e2e85f3e593c8b8 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/cf/3ff581bafb0b4a758e2e85f3e593c8b8 2025-01-08T02:22:26,126 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/5330a89c83ef4074926d1e1630ba855c.c5e2192bbba02a81bf3003d522b64c9b to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/cf/5330a89c83ef4074926d1e1630ba855c.c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:26,128 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/recovered.edits/8.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b/recovered.edits/8.seqid 2025-01-08T02:22:26,128 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/recovered.edits/8.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f/recovered.edits/8.seqid 2025-01-08T02:22:26,128 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/c5e2192bbba02a81bf3003d522b64c9b 2025-01-08T02:22:26,129 DEBUG [HFileArchiver-13 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/6b078cfa4af4c3d42f8fa42264ae001f 2025-01-08T02:22:26,129 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/recovered.edits/12.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598/recovered.edits/12.seqid 2025-01-08T02:22:26,129 DEBUG [HFileArchiver-11 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion-1/a25e31c91636f9ea9e4f60f60a12d598 2025-01-08T02:22:26,129 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithMergeRegion-1 regions 2025-01-08T02:22:26,131 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=128, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,134 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 1 rows of testtb-testExportFileSystemStateWithMergeRegion-1 from hbase:meta 2025-01-08T02:22:26,136 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithMergeRegion-1' descriptor. 2025-01-08T02:22:26,136 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=128, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,136 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithMergeRegion-1' from region states. 2025-01-08T02:22:26,137 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302946136"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:26,138 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 1 regions from META 2025-01-08T02:22:26,138 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => a25e31c91636f9ea9e4f60f60a12d598, NAME => 'testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598.', STARTKEY => '', ENDKEY => ''}] 2025-01-08T02:22:26,138 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithMergeRegion-1' as deleted. 2025-01-08T02:22:26,138 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion-1","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302946138"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:26,140 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithMergeRegion-1 state from META 2025-01-08T02:22:26,142 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=128, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,143 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=128, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion-1 in 31 msec 2025-01-08T02:22:26,222 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=128 2025-01-08T02:22:26,222 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion-1, procId: 128 completed 2025-01-08T02:22:26,222 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,223 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,224 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=129, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,225 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T02:22:26,226 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302946225"}]},"ts":"1736302946225"} 2025-01-08T02:22:26,227 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=DISABLING in hbase:meta 2025-01-08T02:22:26,236 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithMergeRegion to state=DISABLING 2025-01-08T02:22:26,236 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=130, ppid=129, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion}] 2025-01-08T02:22:26,238 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=131, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=1a942e02b6f510a0af53f37e83ed048b, UNASSIGN}, {pid=132, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=b0d1552d5ae3c896125d7ba16ec8a867, UNASSIGN}] 2025-01-08T02:22:26,238 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=132, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=b0d1552d5ae3c896125d7ba16ec8a867, UNASSIGN 2025-01-08T02:22:26,238 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=131, ppid=130, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=1a942e02b6f510a0af53f37e83ed048b, UNASSIGN 2025-01-08T02:22:26,239 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=132 updating hbase:meta row=b0d1552d5ae3c896125d7ba16ec8a867, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:26,239 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=131 updating hbase:meta row=1a942e02b6f510a0af53f37e83ed048b, regionState=CLOSING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:26,240 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:26,240 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=133, ppid=132, state=RUNNABLE; CloseRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:26,241 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:26,241 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=134, ppid=131, state=RUNNABLE; CloseRegionProcedure 1a942e02b6f510a0af53f37e83ed048b, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:26,326 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T02:22:26,391 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:26,391 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(124): Close b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:26,391 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:26,392 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1681): Closing b0d1552d5ae3c896125d7ba16ec8a867, disabling compactions & flushes 2025-01-08T02:22:26,392 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:26,392 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:26,392 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. after waiting 0 ms 2025-01-08T02:22:26,392 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:26,392 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:26,392 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(124): Close 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:26,392 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:26,393 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1681): Closing 1a942e02b6f510a0af53f37e83ed048b, disabling compactions & flushes 2025-01-08T02:22:26,393 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:26,393 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:26,393 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. after waiting 0 ms 2025-01-08T02:22:26,393 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:26,396 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:26,396 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:26,396 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867. 2025-01-08T02:22:26,396 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] regionserver.HRegion(1635): Region close journal for b0d1552d5ae3c896125d7ba16ec8a867: 2025-01-08T02:22:26,397 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:26,398 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:26,398 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b. 2025-01-08T02:22:26,398 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] regionserver.HRegion(1635): Region close journal for 1a942e02b6f510a0af53f37e83ed048b: 2025-01-08T02:22:26,398 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=133}] handler.UnassignRegionHandler(170): Closed b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:26,398 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=132 updating hbase:meta row=b0d1552d5ae3c896125d7ba16ec8a867, regionState=CLOSED 2025-01-08T02:22:26,399 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=134}] handler.UnassignRegionHandler(170): Closed 1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:26,399 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=131 updating hbase:meta row=1a942e02b6f510a0af53f37e83ed048b, regionState=CLOSED 2025-01-08T02:22:26,401 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=133, resume processing ppid=132 2025-01-08T02:22:26,401 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=133, ppid=132, state=SUCCESS; CloseRegionProcedure b0d1552d5ae3c896125d7ba16ec8a867, server=c4e5d52b9759,39911,1736302808913 in 160 msec 2025-01-08T02:22:26,402 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=134, resume processing ppid=131 2025-01-08T02:22:26,402 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=134, ppid=131, state=SUCCESS; CloseRegionProcedure 1a942e02b6f510a0af53f37e83ed048b, server=c4e5d52b9759,37031,1736302808764 in 159 msec 2025-01-08T02:22:26,402 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=132, ppid=130, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=b0d1552d5ae3c896125d7ba16ec8a867, UNASSIGN in 164 msec 2025-01-08T02:22:26,403 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=131, resume processing ppid=130 2025-01-08T02:22:26,403 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=131, ppid=130, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithMergeRegion, region=1a942e02b6f510a0af53f37e83ed048b, UNASSIGN in 165 msec 2025-01-08T02:22:26,405 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=130, resume processing ppid=129 2025-01-08T02:22:26,405 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=130, ppid=129, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 167 msec 2025-01-08T02:22:26,406 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302946406"}]},"ts":"1736302946406"} 2025-01-08T02:22:26,407 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithMergeRegion, state=DISABLED in hbase:meta 2025-01-08T02:22:26,408 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithMergeRegion to state=DISABLED 2025-01-08T02:22:26,410 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=129, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 186 msec 2025-01-08T02:22:26,527 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=129 2025-01-08T02:22:26,527 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 129 completed 2025-01-08T02:22:26,528 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,528 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=135, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,529 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=135, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,529 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,530 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=135, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,531 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,533 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:26,533 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:26,535 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,535 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,535 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,535 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,535 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T02:22:26,535 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/recovered.edits] 2025-01-08T02:22:26,535 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T02:22:26,535 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/recovered.edits] 2025-01-08T02:22:26,535 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T02:22:26,536 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithMergeRegion with data PBUF 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,537 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:26,538 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=135 2025-01-08T02:22:26,540 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/cf/4e36297119b449e699797745b67d736b to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/cf/4e36297119b449e699797745b67d736b 2025-01-08T02:22:26,541 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/cf/74d76542c55e41ae8775c468ab4c4f79 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/cf/74d76542c55e41ae8775c468ab4c4f79 2025-01-08T02:22:26,543 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867/recovered.edits/9.seqid 2025-01-08T02:22:26,543 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b/recovered.edits/9.seqid 2025-01-08T02:22:26,544 DEBUG [HFileArchiver-14 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:26,544 DEBUG [HFileArchiver-15 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithMergeRegion/b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:26,544 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithMergeRegion regions 2025-01-08T02:22:26,544 DEBUG [PEWorker-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c 2025-01-08T02:22:26,545 DEBUG [PEWorker-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf] 2025-01-08T02:22:26,548 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010862f0679b60e7487298d845a3ac712400_b0d1552d5ae3c896125d7ba16ec8a867 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/c4ca4238a0b923820dcc509a6f75849b2025010862f0679b60e7487298d845a3ac712400_b0d1552d5ae3c896125d7ba16ec8a867 2025-01-08T02:22:26,549 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108b6ba1d2b20b94ec98c04e318baf29791_1a942e02b6f510a0af53f37e83ed048b to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c/cf/d41d8cd98f00b204e9800998ecf8427e20250108b6ba1d2b20b94ec98c04e318baf29791_1a942e02b6f510a0af53f37e83ed048b 2025-01-08T02:22:26,550 DEBUG [PEWorker-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithMergeRegion/aaca68e56e63a701f5c1ec9dc2f0511c 2025-01-08T02:22:26,552 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=135, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,554 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemStateWithMergeRegion from hbase:meta 2025-01-08T02:22:26,556 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithMergeRegion' descriptor. 2025-01-08T02:22:26,557 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=135, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,557 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithMergeRegion' from region states. 2025-01-08T02:22:26,557 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302946557"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:26,557 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302946557"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:26,559 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:22:26,559 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 1a942e02b6f510a0af53f37e83ed048b, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,,1736302925961.1a942e02b6f510a0af53f37e83ed048b.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => b0d1552d5ae3c896125d7ba16ec8a867, NAME => 'testtb-testExportFileSystemStateWithMergeRegion,1,1736302925961.b0d1552d5ae3c896125d7ba16ec8a867.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:22:26,559 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithMergeRegion' as deleted. 2025-01-08T02:22:26,559 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithMergeRegion","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302946559"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:26,560 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithMergeRegion state from META 2025-01-08T02:22:26,562 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=135, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,563 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=135, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithMergeRegion in 34 msec 2025-01-08T02:22:26,639 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=135 2025-01-08T02:22:26,640 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithMergeRegion, procId: 135 completed 2025-01-08T02:22:26,648 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportFileSystemStateWithMergeRegion" 2025-01-08T02:22:26,649 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,651 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemStateWithMergeRegion" 2025-01-08T02:22:26,652 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:26,653 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemStateWithMergeRegion-1" 2025-01-08T02:22:26,655 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:26,681 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithMergeRegion Thread=801 (was 784) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-14 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-13 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-28 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:45323 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-30 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-15 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-5 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-29 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-31 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 7368) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-4845 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1081170482_1 at /127.0.0.1:58128 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:45323 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-6 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RPCClient-NioEventLoopGroup-6-7 java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:883) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:58318 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-33 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-32 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #10 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:58278 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #9 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ForkJoinPool.commonPool-worker-4 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Potentially hanging thread: HFileArchiver-11 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: RS_COMPACTED_FILES_DISCHARGER-regionserver/c4e5d52b9759:0-0 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1081170482_1 at /127.0.0.1:51322 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:55900 [Waiting for operation #7] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=809 (was 803) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=668 (was 780), ProcessCount=17 (was 20), AvailableMemoryMB=2184 (was 2039) - AvailableMemoryMB LEAK? - 2025-01-08T02:22:26,681 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=801 is superior to 500 2025-01-08T02:22:26,699 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportExpiredSnapshot Thread=801, OpenFileDescriptor=809, MaxFileDescriptor=1048576, SystemLoadAverage=668, ProcessCount=17, AvailableMemoryMB=2183 2025-01-08T02:22:26,699 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=801 is superior to 500 2025-01-08T02:22:26,700 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:22:26,701 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=136, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:26,702 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:22:26,702 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportExpiredSnapshot" procId is: 136 2025-01-08T02:22:26,703 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:22:26,703 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T02:22:26,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742193_1369 (size=443) 2025-01-08T02:22:26,709 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742193_1369 (size=443) 2025-01-08T02:22:26,710 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742193_1369 (size=443) 2025-01-08T02:22:26,711 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => f80872b7eb7ebb6d6d7c87636cc1f232, NAME => 'testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:26,712 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 8235ae7ab56bd206f032e5c60426552e, NAME => 'testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:26,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742194_1370 (size=68) 2025-01-08T02:22:26,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742194_1370 (size=68) 2025-01-08T02:22:26,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742194_1370 (size=68) 2025-01-08T02:22:26,721 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742195_1371 (size=68) 2025-01-08T02:22:26,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742195_1371 (size=68) 2025-01-08T02:22:26,722 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742195_1371 (size=68) 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1681): Closing 8235ae7ab56bd206f032e5c60426552e, disabling compactions & flushes 2025-01-08T02:22:26,722 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. after waiting 0 ms 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1681): Closing f80872b7eb7ebb6d6d7c87636cc1f232, disabling compactions & flushes 2025-01-08T02:22:26,722 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:26,722 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1635): Region close journal for 8235ae7ab56bd206f032e5c60426552e: 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. after waiting 0 ms 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:26,722 INFO [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:26,722 DEBUG [RegionOpenAndInit-testtb-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1635): Region close journal for f80872b7eb7ebb6d6d7c87636cc1f232: 2025-01-08T02:22:26,723 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:22:26,724 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736302946723"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302946723"}]},"ts":"1736302946723"} 2025-01-08T02:22:26,724 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.","families":{"info":[{"qualifier":"regioninfo","vlen":67,"tag":[],"timestamp":"1736302946723"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302946723"}]},"ts":"1736302946723"} 2025-01-08T02:22:26,725 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:22:26,726 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:22:26,726 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302946726"}]},"ts":"1736302946726"} 2025-01-08T02:22:26,727 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=ENABLING in hbase:meta 2025-01-08T02:22:26,731 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:22:26,732 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:22:26,732 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:22:26,732 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:22:26,732 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:22:26,732 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:22:26,732 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:22:26,732 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:22:26,733 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=f80872b7eb7ebb6d6d7c87636cc1f232, ASSIGN}, {pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8235ae7ab56bd206f032e5c60426552e, ASSIGN}] 2025-01-08T02:22:26,733 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8235ae7ab56bd206f032e5c60426552e, ASSIGN 2025-01-08T02:22:26,733 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=f80872b7eb7ebb6d6d7c87636cc1f232, ASSIGN 2025-01-08T02:22:26,734 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=138, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8235ae7ab56bd206f032e5c60426552e, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:22:26,734 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=137, ppid=136, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=f80872b7eb7ebb6d6d7c87636cc1f232, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:22:26,804 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T02:22:26,884 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:22:26,885 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=137 updating hbase:meta row=f80872b7eb7ebb6d6d7c87636cc1f232, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:26,885 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=138 updating hbase:meta row=8235ae7ab56bd206f032e5c60426552e, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:26,886 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=139, ppid=137, state=RUNNABLE; OpenRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:26,887 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=140, ppid=138, state=RUNNABLE; OpenRegionProcedure 8235ae7ab56bd206f032e5c60426552e, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:22:27,005 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T02:22:27,038 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:27,038 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:27,042 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] handler.AssignRegionHandler(135): Open testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,042 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7285): Opening region: {ENCODED => f80872b7eb7ebb6d6d7c87636cc1f232, NAME => 'testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:22:27,042 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. service=AccessControlService 2025-01-08T02:22:27,043 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:27,043 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportExpiredSnapshot f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,043 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:27,043 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7327): checking encryption for f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,043 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(7330): checking classloading for f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,044 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] handler.AssignRegionHandler(135): Open testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:27,044 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7285): Opening region: {ENCODED => 8235ae7ab56bd206f032e5c60426552e, NAME => 'testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:22:27,044 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. service=AccessControlService 2025-01-08T02:22:27,044 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:27,044 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportExpiredSnapshot 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,044 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(894): Instantiated testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:27,044 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7327): checking encryption for 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,044 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(7330): checking classloading for 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,045 INFO [StoreOpener-f80872b7eb7ebb6d6d7c87636cc1f232-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,045 INFO [StoreOpener-8235ae7ab56bd206f032e5c60426552e-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,046 INFO [StoreOpener-f80872b7eb7ebb6d6d7c87636cc1f232-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region f80872b7eb7ebb6d6d7c87636cc1f232 columnFamilyName cf 2025-01-08T02:22:27,047 INFO [StoreOpener-8235ae7ab56bd206f032e5c60426552e-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 8235ae7ab56bd206f032e5c60426552e columnFamilyName cf 2025-01-08T02:22:27,047 DEBUG [StoreOpener-8235ae7ab56bd206f032e5c60426552e-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:27,047 DEBUG [StoreOpener-f80872b7eb7ebb6d6d7c87636cc1f232-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:27,048 INFO [StoreOpener-8235ae7ab56bd206f032e5c60426552e-1 {}] regionserver.HStore(327): Store=8235ae7ab56bd206f032e5c60426552e/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:27,048 INFO [StoreOpener-f80872b7eb7ebb6d6d7c87636cc1f232-1 {}] regionserver.HStore(327): Store=f80872b7eb7ebb6d6d7c87636cc1f232/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:27,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,051 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1085): writing seq id for 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,051 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1085): writing seq id for f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,053 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:27,053 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:27,053 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1102): Opened f80872b7eb7ebb6d6d7c87636cc1f232; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68575290, jitterRate=0.021851450204849243}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:27,053 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1102): Opened 8235ae7ab56bd206f032e5c60426552e; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=68644777, jitterRate=0.022886887192726135}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:27,054 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegion(1001): Region open journal for f80872b7eb7ebb6d6d7c87636cc1f232: 2025-01-08T02:22:27,054 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegion(1001): Region open journal for 8235ae7ab56bd206f032e5c60426552e: 2025-01-08T02:22:27,055 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e., pid=140, masterSystemTime=1736302947038 2025-01-08T02:22:27,055 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232., pid=139, masterSystemTime=1736302947038 2025-01-08T02:22:27,056 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:27,056 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=140}] handler.AssignRegionHandler(164): Opened testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:27,056 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=138 updating hbase:meta row=8235ae7ab56bd206f032e5c60426552e, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:27,057 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,057 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=139}] handler.AssignRegionHandler(164): Opened testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,057 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=137 updating hbase:meta row=f80872b7eb7ebb6d6d7c87636cc1f232, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:27,059 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=140, resume processing ppid=138 2025-01-08T02:22:27,059 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=140, ppid=138, state=SUCCESS; OpenRegionProcedure 8235ae7ab56bd206f032e5c60426552e, server=c4e5d52b9759,36743,1736302808837 in 171 msec 2025-01-08T02:22:27,060 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=139, resume processing ppid=137 2025-01-08T02:22:27,060 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=138, ppid=136, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8235ae7ab56bd206f032e5c60426552e, ASSIGN in 328 msec 2025-01-08T02:22:27,061 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=139, ppid=137, state=SUCCESS; OpenRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232, server=c4e5d52b9759,39911,1736302808913 in 172 msec 2025-01-08T02:22:27,062 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=137, resume processing ppid=136 2025-01-08T02:22:27,062 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=137, ppid=136, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=f80872b7eb7ebb6d6d7c87636cc1f232, ASSIGN in 328 msec 2025-01-08T02:22:27,062 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:22:27,062 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302947062"}]},"ts":"1736302947062"} 2025-01-08T02:22:27,063 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=ENABLED in hbase:meta 2025-01-08T02:22:27,066 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=136, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportExpiredSnapshot execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:22:27,066 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportExpiredSnapshot jenkins: RWXCA 2025-01-08T02:22:27,067 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T02:22:27,069 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:27,069 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:27,069 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:27,069 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:27,071 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:27,071 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:27,071 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:27,071 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:27,072 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=136, state=SUCCESS; CreateTableProcedure table=testtb-testExportExpiredSnapshot in 371 msec 2025-01-08T02:22:27,306 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=136 2025-01-08T02:22:27,307 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 136 completed 2025-01-08T02:22:27,310 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportExpiredSnapshot 2025-01-08T02:22:27,310 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,310 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:27,327 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T02:22:27,327 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302947327 (current time:1736302947327). 2025-01-08T02:22:27,327 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:27,327 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T02:22:27,327 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:27,328 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x036ced7b to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@63989abb 2025-01-08T02:22:27,332 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@43c9853d, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:27,334 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:27,335 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59736, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:27,336 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x036ced7b to 127.0.0.1:61593 2025-01-08T02:22:27,337 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:27,338 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x62dc7601 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@45b675a3 2025-01-08T02:22:27,341 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@4a59f045, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:27,349 DEBUG [hconnection-0x363ee612-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:27,350 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59746, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:27,352 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:27,353 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49698, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:27,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x62dc7601 to 127.0.0.1:61593 2025-01-08T02:22:27,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:27,354 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T02:22:27,355 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:27,356 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=141, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T02:22:27,356 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 141 2025-01-08T02:22:27,357 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:27,357 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T02:22:27,358 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:27,360 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:27,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742196_1372 (size=170) 2025-01-08T02:22:27,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742196_1372 (size=170) 2025-01-08T02:22:27,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742196_1372 (size=170) 2025-01-08T02:22:27,373 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:27,373 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232}, {pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e}] 2025-01-08T02:22:27,374 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,374 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,458 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T02:22:27,525 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:27,525 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:27,526 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=143 2025-01-08T02:22:27,526 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=142 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.HRegion(2538): Flush status journal for 8235ae7ab56bd206f032e5c60426552e: 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.HRegion(2538): Flush status journal for f80872b7eb7ebb6d6d7c87636cc1f232: 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. for emptySnaptb0-testExportExpiredSnapshot completed. 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. for emptySnaptb0-testExportExpiredSnapshot completed. 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.' region-info for snapshot=emptySnaptb0-testExportExpiredSnapshot 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.' region-info for snapshot=emptySnaptb0-testExportExpiredSnapshot 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:27,526 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:27,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742197_1373 (size=71) 2025-01-08T02:22:27,533 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742197_1373 (size=71) 2025-01-08T02:22:27,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742197_1373 (size=71) 2025-01-08T02:22:27,534 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:27,534 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=143}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=143 2025-01-08T02:22:27,534 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742198_1374 (size=71) 2025-01-08T02:22:27,534 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=143 2025-01-08T02:22:27,535 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportExpiredSnapshot on region 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742198_1374 (size=71) 2025-01-08T02:22:27,535 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=143, ppid=141, state=RUNNABLE; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,535 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742198_1374 (size=71) 2025-01-08T02:22:27,536 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,536 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=142}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=142 2025-01-08T02:22:27,536 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=142 2025-01-08T02:22:27,536 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportExpiredSnapshot on region f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,537 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=142, ppid=141, state=RUNNABLE; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,537 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=143, ppid=141, state=SUCCESS; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e in 163 msec 2025-01-08T02:22:27,538 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=142, resume processing ppid=141 2025-01-08T02:22:27,538 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=142, ppid=141, state=SUCCESS; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232 in 164 msec 2025-01-08T02:22:27,538 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:27,539 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:27,540 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:27,540 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:27,540 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:27,540 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:22:27,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742199_1375 (size=63) 2025-01-08T02:22:27,551 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742199_1375 (size=63) 2025-01-08T02:22:27,552 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742199_1375 (size=63) 2025-01-08T02:22:27,553 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:27,553 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportExpiredSnapshot 2025-01-08T02:22:27,553 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportExpiredSnapshot 2025-01-08T02:22:27,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742200_1376 (size=653) 2025-01-08T02:22:27,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742200_1376 (size=653) 2025-01-08T02:22:27,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742200_1376 (size=653) 2025-01-08T02:22:27,569 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:27,573 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:27,573 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportExpiredSnapshot to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testExportExpiredSnapshot 2025-01-08T02:22:27,574 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=141, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:27,574 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 141 2025-01-08T02:22:27,576 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=141, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=141, snapshot={ ss=emptySnaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } in 219 msec 2025-01-08T02:22:27,659 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=141 2025-01-08T02:22:27,660 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportExpiredSnapshot, procId: 141 completed 2025-01-08T02:22:27,662 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:27,663 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.HRegion(8254): writing data to region testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:27,666 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportExpiredSnapshot 2025-01-08T02:22:27,666 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,666 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:27,675 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T02:22:27,676 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302947676 (current time:1736302947676). 2025-01-08T02:22:27,676 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:27,676 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T02:22:27,676 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:27,677 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x44608924 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@52ff9cc 2025-01-08T02:22:27,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2d418d71, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:27,681 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:27,682 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59762, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:27,682 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x44608924 to 127.0.0.1:61593 2025-01-08T02:22:27,683 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:27,683 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x58b7c3b3 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@222ca0f0 2025-01-08T02:22:27,686 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3a79bc1, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:27,687 DEBUG [hconnection-0x6992bce8-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:27,688 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59766, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:27,689 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:27,690 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49704, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:27,690 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x58b7c3b3 to 127.0.0.1:61593 2025-01-08T02:22:27,690 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:27,690 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T02:22:27,691 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:27,692 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=144, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } 2025-01-08T02:22:27,692 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 144 2025-01-08T02:22:27,693 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:27,693 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T02:22:27,693 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:27,696 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:27,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742201_1377 (size=165) 2025-01-08T02:22:27,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742201_1377 (size=165) 2025-01-08T02:22:27,701 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742201_1377 (size=165) 2025-01-08T02:22:27,702 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:27,702 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232}, {pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e}] 2025-01-08T02:22:27,703 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,703 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,794 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T02:22:27,854 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:27,854 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:27,855 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=146 2025-01-08T02:22:27,855 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=145 2025-01-08T02:22:27,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:27,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,855 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(2837): Flushing f80872b7eb7ebb6d6d7c87636cc1f232 1/1 column families, dataSize=333 B heapSize=976 B 2025-01-08T02:22:27,855 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(2837): Flushing 8235ae7ab56bd206f032e5c60426552e 1/1 column families, dataSize=2.93 KB heapSize=6.58 KB 2025-01-08T02:22:27,875 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108f6f240ee10b84b798e43d1ad97e9a8bb_f80872b7eb7ebb6d6d7c87636cc1f232 is 71, key is 00e3905003ea74610bd566496b0de94a/cf:q/1736302947662/Put/seqid=0 2025-01-08T02:22:27,881 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010877725646de9f49228a1405aab0279f12_8235ae7ab56bd206f032e5c60426552e is 71, key is 11d913f4c67ae6e1dd73c2ba72f2367a/cf:q/1736302947663/Put/seqid=0 2025-01-08T02:22:27,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742202_1378 (size=5242) 2025-01-08T02:22:27,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742202_1378 (size=5242) 2025-01-08T02:22:27,884 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742202_1378 (size=5242) 2025-01-08T02:22:27,885 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:27,889 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108f6f240ee10b84b798e43d1ad97e9a8bb_f80872b7eb7ebb6d6d7c87636cc1f232 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e20250108f6f240ee10b84b798e43d1ad97e9a8bb_f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742203_1379 (size=8031) 2025-01-08T02:22:27,890 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/.tmp/cf/ef109958ed694211b2962ffb83afeb47, store: [table=testtb-testExportExpiredSnapshot family=cf region=f80872b7eb7ebb6d6d7c87636cc1f232] 2025-01-08T02:22:27,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742203_1379 (size=8031) 2025-01-08T02:22:27,891 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/.tmp/cf/ef109958ed694211b2962ffb83afeb47 is 209, key is 0c6dd7740fba04e12c1513ab34e963c70/cf:q/1736302947662/Put/seqid=0 2025-01-08T02:22:27,891 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742203_1379 (size=8031) 2025-01-08T02:22:27,891 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:27,896 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010877725646de9f49228a1405aab0279f12_8235ae7ab56bd206f032e5c60426552e to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b2025010877725646de9f49228a1405aab0279f12_8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:27,896 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742204_1380 (size=6328) 2025-01-08T02:22:27,897 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/.tmp/cf/fcf21c0566d44aa78ae187a4bc31e4b3, store: [table=testtb-testExportExpiredSnapshot family=cf region=8235ae7ab56bd206f032e5c60426552e] 2025-01-08T02:22:27,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742204_1380 (size=6328) 2025-01-08T02:22:27,898 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/.tmp/cf/fcf21c0566d44aa78ae187a4bc31e4b3 is 209, key is 1f63a0610d16dada4a131e2d687bc4761/cf:q/1736302947663/Put/seqid=0 2025-01-08T02:22:27,898 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742204_1380 (size=6328) 2025-01-08T02:22:27,898 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=333, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/.tmp/cf/ef109958ed694211b2962ffb83afeb47 2025-01-08T02:22:27,905 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/.tmp/cf/ef109958ed694211b2962ffb83afeb47 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/cf/ef109958ed694211b2962ffb83afeb47 2025-01-08T02:22:27,914 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742205_1381 (size=14587) 2025-01-08T02:22:27,915 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742205_1381 (size=14587) 2025-01-08T02:22:27,916 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/cf/ef109958ed694211b2962ffb83afeb47, entries=5, sequenceid=6, filesize=6.2 K 2025-01-08T02:22:27,916 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742205_1381 (size=14587) 2025-01-08T02:22:27,917 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(3040): Finished flush of dataSize ~333 B/333, heapSize ~960 B/960, currentSize=0 B/0 for f80872b7eb7ebb6d6d7c87636cc1f232 in 62ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:27,917 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportExpiredSnapshot' 2025-01-08T02:22:27,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.HRegion(2538): Flush status journal for f80872b7eb7ebb6d6d7c87636cc1f232: 2025-01-08T02:22:27,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. for snaptb0-testExportExpiredSnapshot completed. 2025-01-08T02:22:27,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.' region-info for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:27,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:27,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/cf/ef109958ed694211b2962ffb83afeb47] hfiles 2025-01-08T02:22:27,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/cf/ef109958ed694211b2962ffb83afeb47 for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:27,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742206_1382 (size=110) 2025-01-08T02:22:27,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742206_1382 (size=110) 2025-01-08T02:22:27,937 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742206_1382 (size=110) 2025-01-08T02:22:27,939 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:27,939 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=145}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=145 2025-01-08T02:22:27,939 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=145 2025-01-08T02:22:27,939 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportExpiredSnapshot on region f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,939 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=145, ppid=144, state=RUNNABLE; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:27,941 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=145, ppid=144, state=SUCCESS; SnapshotRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232 in 238 msec 2025-01-08T02:22:27,995 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T02:22:28,296 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T02:22:28,315 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=2.9 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/.tmp/cf/fcf21c0566d44aa78ae187a4bc31e4b3 2025-01-08T02:22:28,321 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/.tmp/cf/fcf21c0566d44aa78ae187a4bc31e4b3 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/cf/fcf21c0566d44aa78ae187a4bc31e4b3 2025-01-08T02:22:28,325 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/cf/fcf21c0566d44aa78ae187a4bc31e4b3, entries=45, sequenceid=6, filesize=14.2 K 2025-01-08T02:22:28,326 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(3040): Finished flush of dataSize ~2.93 KB/3003, heapSize ~6.56 KB/6720, currentSize=0 B/0 for 8235ae7ab56bd206f032e5c60426552e in 471ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:28,326 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.HRegion(2538): Flush status journal for 8235ae7ab56bd206f032e5c60426552e: 2025-01-08T02:22:28,326 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. for snaptb0-testExportExpiredSnapshot completed. 2025-01-08T02:22:28,327 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.' region-info for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:28,327 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:28,327 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/cf/fcf21c0566d44aa78ae187a4bc31e4b3] hfiles 2025-01-08T02:22:28,327 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/cf/fcf21c0566d44aa78ae187a4bc31e4b3 for snapshot=snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:28,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742207_1383 (size=110) 2025-01-08T02:22:28,332 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742207_1383 (size=110) 2025-01-08T02:22:28,333 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742207_1383 (size=110) 2025-01-08T02:22:28,333 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:28,333 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=146}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=146 2025-01-08T02:22:28,333 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=146 2025-01-08T02:22:28,333 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportExpiredSnapshot on region 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:28,333 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=146, ppid=144, state=RUNNABLE; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:28,336 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:28,336 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=146, resume processing ppid=144 2025-01-08T02:22:28,337 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=146, ppid=144, state=SUCCESS; SnapshotRegionProcedure 8235ae7ab56bd206f032e5c60426552e in 632 msec 2025-01-08T02:22:28,337 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:28,338 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:28,338 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:28,338 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:28,341 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b2025010877725646de9f49228a1405aab0279f12_8235ae7ab56bd206f032e5c60426552e, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e20250108f6f240ee10b84b798e43d1ad97e9a8bb_f80872b7eb7ebb6d6d7c87636cc1f232] hfiles 2025-01-08T02:22:28,341 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b2025010877725646de9f49228a1405aab0279f12_8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:28,341 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e20250108f6f240ee10b84b798e43d1ad97e9a8bb_f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:28,352 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742208_1384 (size=294) 2025-01-08T02:22:28,352 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742208_1384 (size=294) 2025-01-08T02:22:28,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742208_1384 (size=294) 2025-01-08T02:22:28,354 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:28,354 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:28,355 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:28,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742209_1385 (size=963) 2025-01-08T02:22:28,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742209_1385 (size=963) 2025-01-08T02:22:28,373 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742209_1385 (size=963) 2025-01-08T02:22:28,375 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:28,380 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:28,380 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportExpiredSnapshot to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:28,381 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=144, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:28,381 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 }, snapshot procedure id = 144 2025-01-08T02:22:28,382 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=144, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=144, snapshot={ ss=snaptb0-testExportExpiredSnapshot table=testtb-testExportExpiredSnapshot type=FLUSH ttl=0 } in 690 msec 2025-01-08T02:22:28,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot 2025-01-08T02:22:28,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot Metrics about Tables on a single HBase RegionServer 2025-01-08T02:22:28,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion-1 2025-01-08T02:22:28,433 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithMergeRegion 2025-01-08T02:22:28,798 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=144 2025-01-08T02:22:28,798 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportExpiredSnapshot, procId: 144 completed 2025-01-08T02:22:28,799 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:22:28,800 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=147, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testExportExpiredSnapshot 2025-01-08T02:22:28,801 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:22:28,801 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testExportExpiredSnapshot" procId is: 147 2025-01-08T02:22:28,802 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:22:28,802 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T02:22:28,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742210_1386 (size=436) 2025-01-08T02:22:28,808 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742210_1386 (size=436) 2025-01-08T02:22:28,809 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742210_1386 (size=436) 2025-01-08T02:22:28,810 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 2f4301e371bc30413722a847ac142fb1, NAME => 'testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:28,811 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 4c7db76a9413d28afb5a4d977c0db552, NAME => 'testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testExportExpiredSnapshot', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:28,818 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742212_1388 (size=61) 2025-01-08T02:22:28,818 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742212_1388 (size=61) 2025-01-08T02:22:28,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742212_1388 (size=61) 2025-01-08T02:22:28,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742211_1387 (size=61) 2025-01-08T02:22:28,820 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:28,820 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1681): Closing 4c7db76a9413d28afb5a4d977c0db552, disabling compactions & flushes 2025-01-08T02:22:28,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742211_1387 (size=61) 2025-01-08T02:22:28,820 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:28,820 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:28,820 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. after waiting 0 ms 2025-01-08T02:22:28,820 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:28,820 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:28,820 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-1 {}] regionserver.HRegion(1635): Region close journal for 4c7db76a9413d28afb5a4d977c0db552: 2025-01-08T02:22:28,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742211_1387 (size=61) 2025-01-08T02:22:28,831 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:28,831 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1681): Closing 2f4301e371bc30413722a847ac142fb1, disabling compactions & flushes 2025-01-08T02:22:28,831 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:28,831 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:28,831 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. after waiting 0 ms 2025-01-08T02:22:28,831 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:28,831 INFO [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:28,831 DEBUG [RegionOpenAndInit-testExportExpiredSnapshot-pool-0 {}] regionserver.HRegion(1635): Region close journal for 2f4301e371bc30413722a847ac142fb1: 2025-01-08T02:22:28,836 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:22:28,836 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552.","families":{"info":[{"qualifier":"regioninfo","vlen":60,"tag":[],"timestamp":"1736302948836"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302948836"}]},"ts":"1736302948836"} 2025-01-08T02:22:28,836 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1.","families":{"info":[{"qualifier":"regioninfo","vlen":60,"tag":[],"timestamp":"1736302948836"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302948836"}]},"ts":"1736302948836"} 2025-01-08T02:22:28,839 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:22:28,839 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:22:28,840 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302948839"}]},"ts":"1736302948839"} 2025-01-08T02:22:28,841 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportExpiredSnapshot, state=ENABLING in hbase:meta 2025-01-08T02:22:28,845 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:22:28,846 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:22:28,846 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:22:28,846 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:22:28,846 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:22:28,846 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:22:28,846 INFO [PEWorker-2 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:22:28,846 DEBUG [PEWorker-2 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:22:28,847 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=2f4301e371bc30413722a847ac142fb1, ASSIGN}, {pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=4c7db76a9413d28afb5a4d977c0db552, ASSIGN}] 2025-01-08T02:22:28,848 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=4c7db76a9413d28afb5a4d977c0db552, ASSIGN 2025-01-08T02:22:28,848 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=2f4301e371bc30413722a847ac142fb1, ASSIGN 2025-01-08T02:22:28,848 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=149, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=4c7db76a9413d28afb5a4d977c0db552, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:22:28,848 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=148, ppid=147, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=2f4301e371bc30413722a847ac142fb1, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:22:28,903 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T02:22:28,999 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:22:28,999 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=149 updating hbase:meta row=4c7db76a9413d28afb5a4d977c0db552, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:28,999 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=148 updating hbase:meta row=2f4301e371bc30413722a847ac142fb1, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:29,001 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=150, ppid=149, state=RUNNABLE; OpenRegionProcedure 4c7db76a9413d28afb5a4d977c0db552, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:29,002 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=151, ppid=148, state=RUNNABLE; OpenRegionProcedure 2f4301e371bc30413722a847ac142fb1, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:29,104 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T02:22:29,153 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:29,154 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:29,156 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] handler.AssignRegionHandler(135): Open testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:29,156 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7285): Opening region: {ENCODED => 4c7db76a9413d28afb5a4d977c0db552, NAME => 'testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:22:29,156 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] handler.AssignRegionHandler(135): Open testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:29,156 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7285): Opening region: {ENCODED => 2f4301e371bc30413722a847ac142fb1, NAME => 'testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. service=AccessControlService 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7999): Registered coprocessor service: region=testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. service=AccessControlService 2025-01-08T02:22:29,157 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:29,157 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportExpiredSnapshot 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testExportExpiredSnapshot 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(894): Instantiated testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7327): checking encryption for 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(7330): checking classloading for 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7327): checking encryption for 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,157 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(7330): checking classloading for 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,159 INFO [StoreOpener-2f4301e371bc30413722a847ac142fb1-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,159 INFO [StoreOpener-4c7db76a9413d28afb5a4d977c0db552-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,163 INFO [StoreOpener-2f4301e371bc30413722a847ac142fb1-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 2f4301e371bc30413722a847ac142fb1 columnFamilyName cf 2025-01-08T02:22:29,163 INFO [StoreOpener-4c7db76a9413d28afb5a4d977c0db552-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 4c7db76a9413d28afb5a4d977c0db552 columnFamilyName cf 2025-01-08T02:22:29,164 DEBUG [StoreOpener-4c7db76a9413d28afb5a4d977c0db552-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:29,164 DEBUG [StoreOpener-2f4301e371bc30413722a847ac142fb1-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:29,164 INFO [StoreOpener-4c7db76a9413d28afb5a4d977c0db552-1 {}] regionserver.HStore(327): Store=4c7db76a9413d28afb5a4d977c0db552/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:29,164 INFO [StoreOpener-2f4301e371bc30413722a847ac142fb1-1 {}] regionserver.HStore(327): Store=2f4301e371bc30413722a847ac142fb1/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:29,165 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,165 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,165 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,165 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,167 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1085): writing seq id for 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,167 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1085): writing seq id for 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,169 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:29,169 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:29,170 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1102): Opened 2f4301e371bc30413722a847ac142fb1; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72435001, jitterRate=0.07936562597751617}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:29,170 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1102): Opened 4c7db76a9413d28afb5a4d977c0db552; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=73518188, jitterRate=0.09550637006759644}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:29,171 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegion(1001): Region open journal for 2f4301e371bc30413722a847ac142fb1: 2025-01-08T02:22:29,171 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegion(1001): Region open journal for 4c7db76a9413d28afb5a4d977c0db552: 2025-01-08T02:22:29,171 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552., pid=150, masterSystemTime=1736302949153 2025-01-08T02:22:29,171 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegionServer(2601): Post open deploy tasks for testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1., pid=151, masterSystemTime=1736302949154 2025-01-08T02:22:29,173 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:29,173 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=151}] handler.AssignRegionHandler(164): Opened testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:29,173 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=148 updating hbase:meta row=2f4301e371bc30413722a847ac142fb1, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:29,173 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] regionserver.HRegionServer(2628): Finished post open deploy task for testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:29,173 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=150}] handler.AssignRegionHandler(164): Opened testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:29,174 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=149 updating hbase:meta row=4c7db76a9413d28afb5a4d977c0db552, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:29,176 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=151, resume processing ppid=148 2025-01-08T02:22:29,176 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=151, ppid=148, state=SUCCESS; OpenRegionProcedure 2f4301e371bc30413722a847ac142fb1, server=c4e5d52b9759,37031,1736302808764 in 173 msec 2025-01-08T02:22:29,177 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=150, resume processing ppid=149 2025-01-08T02:22:29,177 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=150, ppid=149, state=SUCCESS; OpenRegionProcedure 4c7db76a9413d28afb5a4d977c0db552, server=c4e5d52b9759,39911,1736302808913 in 174 msec 2025-01-08T02:22:29,178 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=148, ppid=147, state=SUCCESS; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=2f4301e371bc30413722a847ac142fb1, ASSIGN in 329 msec 2025-01-08T02:22:29,179 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=149, resume processing ppid=147 2025-01-08T02:22:29,179 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=149, ppid=147, state=SUCCESS; TransitRegionStateProcedure table=testExportExpiredSnapshot, region=4c7db76a9413d28afb5a4d977c0db552, ASSIGN in 330 msec 2025-01-08T02:22:29,179 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:22:29,180 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302949179"}]},"ts":"1736302949179"} 2025-01-08T02:22:29,181 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testExportExpiredSnapshot, state=ENABLED in hbase:meta 2025-01-08T02:22:29,188 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=147, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testExportExpiredSnapshot execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:22:29,188 DEBUG [PEWorker-4 {}] access.PermissionStorage(175): Writing permission with rowKey testExportExpiredSnapshot jenkins: RWXCA 2025-01-08T02:22:29,190 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T02:22:29,193 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:29,193 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:29,194 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:29,194 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF\x0AF\x0A\x07jenkins\x12;\x08\x03"7\x0A+\x0A\x07default\x12 testtb-testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,198 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:29,201 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=147, state=SUCCESS; CreateTableProcedure table=testExportExpiredSnapshot in 398 msec 2025-01-08T02:22:29,279 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testExportExpiredSnapshot' 2025-01-08T02:22:29,406 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=147 2025-01-08T02:22:29,406 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testExportExpiredSnapshot, procId: 147 completed 2025-01-08T02:22:29,408 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportExpiredSnapshot 2025-01-08T02:22:29,408 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:29,409 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:29,427 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.HRegion(8254): writing data to region testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:29,428 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:29,434 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testExportExpiredSnapshot 2025-01-08T02:22:29,434 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:29,435 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:29,445 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } 2025-01-08T02:22:29,445 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snapshot-testExportExpiredSnapshot VERSION not specified, setting to 2 2025-01-08T02:22:29,445 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:29,446 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x76dfa02a to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@2c675efa 2025-01-08T02:22:29,467 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7e0ba4ea, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:29,473 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:29,474 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59768, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:29,475 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x76dfa02a to 127.0.0.1:61593 2025-01-08T02:22:29,476 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:29,476 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x173488dc to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@132944e 2025-01-08T02:22:29,481 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@22bd5c76, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:29,483 DEBUG [hconnection-0x3361e0ce-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:29,484 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:59776, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:29,486 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:29,487 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:49712, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:29,488 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x173488dc to 127.0.0.1:61593 2025-01-08T02:22:29,488 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:29,488 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testExportExpiredSnapshot], kv [jenkins: RWXCA] 2025-01-08T02:22:29,489 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:29,490 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=152, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } 2025-01-08T02:22:29,490 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 }, snapshot procedure id = 152 2025-01-08T02:22:29,491 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:29,491 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T02:22:29,492 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:29,495 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:29,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742213_1389 (size=152) 2025-01-08T02:22:29,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742213_1389 (size=152) 2025-01-08T02:22:29,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742213_1389 (size=152) 2025-01-08T02:22:29,514 INFO [PEWorker-3 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:29,515 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 2f4301e371bc30413722a847ac142fb1}, {pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 4c7db76a9413d28afb5a4d977c0db552}] 2025-01-08T02:22:29,516 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,516 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,592 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T02:22:29,667 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:29,667 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:29,667 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=154 2025-01-08T02:22:29,668 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:29,668 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=153 2025-01-08T02:22:29,668 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:29,668 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(2837): Flushing 4c7db76a9413d28afb5a4d977c0db552 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T02:22:29,668 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(2837): Flushing 2f4301e371bc30413722a847ac142fb1 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T02:22:29,701 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108a3b79f9f7e5742bdbaae5d60a64e43ca_2f4301e371bc30413722a847ac142fb1 is 71, key is 01bc6b57cc903e1ee444853cb4f0d76e/cf:q/1736302949427/Put/seqid=0 2025-01-08T02:22:29,719 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742214_1390 (size=5171) 2025-01-08T02:22:29,720 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108a398a22ef22d4ef9b9d614bd8d011e19_4c7db76a9413d28afb5a4d977c0db552 is 71, key is 12fd341240d57925209b61a1783edbeb/cf:q/1736302949428/Put/seqid=0 2025-01-08T02:22:29,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742214_1390 (size=5171) 2025-01-08T02:22:29,720 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742214_1390 (size=5171) 2025-01-08T02:22:29,723 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:29,728 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108a3b79f9f7e5742bdbaae5d60a64e43ca_2f4301e371bc30413722a847ac142fb1 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e20250108a3b79f9f7e5742bdbaae5d60a64e43ca_2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,729 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/.tmp/cf/3b39528cc58647fbb6700209ef06f6be, store: [table=testExportExpiredSnapshot family=cf region=2f4301e371bc30413722a847ac142fb1] 2025-01-08T02:22:29,730 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/.tmp/cf/3b39528cc58647fbb6700209ef06f6be is 202, key is 0286c63c1125cbe8b8f91d40e1d30194e/cf:q/1736302949427/Put/seqid=0 2025-01-08T02:22:29,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742215_1391 (size=8101) 2025-01-08T02:22:29,747 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742215_1391 (size=8101) 2025-01-08T02:22:29,748 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742215_1391 (size=8101) 2025-01-08T02:22:29,749 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:29,755 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108a398a22ef22d4ef9b9d614bd8d011e19_4c7db76a9413d28afb5a4d977c0db552 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b20250108a398a22ef22d4ef9b9d614bd8d011e19_4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,757 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/.tmp/cf/a95b136ac56f48d9bdd8ebff597d7079, store: [table=testExportExpiredSnapshot family=cf region=4c7db76a9413d28afb5a4d977c0db552] 2025-01-08T02:22:29,757 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/.tmp/cf/a95b136ac56f48d9bdd8ebff597d7079 is 202, key is 14a08b16781e5e41b13d54b5fd4f830f5/cf:q/1736302949428/Put/seqid=0 2025-01-08T02:22:29,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742216_1392 (size=6086) 2025-01-08T02:22:29,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742216_1392 (size=6086) 2025-01-08T02:22:29,770 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742216_1392 (size=6086) 2025-01-08T02:22:29,771 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/.tmp/cf/3b39528cc58647fbb6700209ef06f6be 2025-01-08T02:22:29,777 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/.tmp/cf/3b39528cc58647fbb6700209ef06f6be as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/cf/3b39528cc58647fbb6700209ef06f6be 2025-01-08T02:22:29,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742217_1393 (size=14463) 2025-01-08T02:22:29,779 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742217_1393 (size=14463) 2025-01-08T02:22:29,780 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742217_1393 (size=14463) 2025-01-08T02:22:29,781 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=5, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/.tmp/cf/a95b136ac56f48d9bdd8ebff597d7079 2025-01-08T02:22:29,784 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/cf/3b39528cc58647fbb6700209ef06f6be, entries=4, sequenceid=5, filesize=5.9 K 2025-01-08T02:22:29,786 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 2f4301e371bc30413722a847ac142fb1 in 118ms, sequenceid=5, compaction requested=false 2025-01-08T02:22:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.HRegion(2538): Flush status journal for 2f4301e371bc30413722a847ac142fb1: 2025-01-08T02:22:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. for snapshot-testExportExpiredSnapshot completed. 2025-01-08T02:22:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(241): Storing 'testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1.' region-info for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T02:22:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/cf/3b39528cc58647fbb6700209ef06f6be] hfiles 2025-01-08T02:22:29,786 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/cf/3b39528cc58647fbb6700209ef06f6be for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T02:22:29,789 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/.tmp/cf/a95b136ac56f48d9bdd8ebff597d7079 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/cf/a95b136ac56f48d9bdd8ebff597d7079 2025-01-08T02:22:29,794 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T02:22:29,794 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/cf/a95b136ac56f48d9bdd8ebff597d7079, entries=46, sequenceid=5, filesize=14.1 K 2025-01-08T02:22:29,796 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for 4c7db76a9413d28afb5a4d977c0db552 in 127ms, sequenceid=5, compaction requested=false 2025-01-08T02:22:29,796 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.HRegion(2538): Flush status journal for 4c7db76a9413d28afb5a4d977c0db552: 2025-01-08T02:22:29,796 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(75): Snapshotting region testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. for snapshot-testExportExpiredSnapshot completed. 2025-01-08T02:22:29,796 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(241): Storing 'testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552.' region-info for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T02:22:29,796 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:29,796 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/cf/a95b136ac56f48d9bdd8ebff597d7079] hfiles 2025-01-08T02:22:29,796 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/cf/a95b136ac56f48d9bdd8ebff597d7079 for snapshot=snapshot-testExportExpiredSnapshot 2025-01-08T02:22:29,817 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742218_1394 (size=103) 2025-01-08T02:22:29,817 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742218_1394 (size=103) 2025-01-08T02:22:29,818 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742218_1394 (size=103) 2025-01-08T02:22:29,819 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:22:29,819 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=153}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=153 2025-01-08T02:22:29,819 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=153 2025-01-08T02:22:29,819 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snapshot-testExportExpiredSnapshot on region 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,820 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=153, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,826 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=153, ppid=152, state=SUCCESS; SnapshotRegionProcedure 2f4301e371bc30413722a847ac142fb1 in 309 msec 2025-01-08T02:22:29,845 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742219_1395 (size=103) 2025-01-08T02:22:29,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742219_1395 (size=103) 2025-01-08T02:22:29,846 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742219_1395 (size=103) 2025-01-08T02:22:29,847 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:22:29,847 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=154}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=154 2025-01-08T02:22:29,847 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=154 2025-01-08T02:22:29,847 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snapshot-testExportExpiredSnapshot on region 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,847 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=154, ppid=152, state=RUNNABLE; SnapshotRegionProcedure 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,851 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=154, resume processing ppid=152 2025-01-08T02:22:29,851 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=154, ppid=152, state=SUCCESS; SnapshotRegionProcedure 4c7db76a9413d28afb5a4d977c0db552 in 333 msec 2025-01-08T02:22:29,851 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:29,852 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:29,853 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:29,853 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:29,853 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:29,854 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b20250108a398a22ef22d4ef9b9d614bd8d011e19_4c7db76a9413d28afb5a4d977c0db552, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e20250108a3b79f9f7e5742bdbaae5d60a64e43ca_2f4301e371bc30413722a847ac142fb1] hfiles 2025-01-08T02:22:29,854 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/c4ca4238a0b923820dcc509a6f75849b20250108a398a22ef22d4ef9b9d614bd8d011e19_4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:22:29,854 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testExportExpiredSnapshot/1030f41967fbb659ab4c2a7a1774d313/cf/d41d8cd98f00b204e9800998ecf8427e20250108a3b79f9f7e5742bdbaae5d60a64e43ca_2f4301e371bc30413722a847ac142fb1 2025-01-08T02:22:29,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742220_1396 (size=287) 2025-01-08T02:22:29,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742220_1396 (size=287) 2025-01-08T02:22:29,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742220_1396 (size=287) 2025-01-08T02:22:29,870 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:29,870 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snapshot-testExportExpiredSnapshot 2025-01-08T02:22:29,871 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot 2025-01-08T02:22:29,903 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742221_1397 (size=935) 2025-01-08T02:22:29,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742221_1397 (size=935) 2025-01-08T02:22:29,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742221_1397 (size=935) 2025-01-08T02:22:29,908 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:29,915 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:29,916 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snapshot-testExportExpiredSnapshot 2025-01-08T02:22:29,917 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=152, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:29,917 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 }, snapshot procedure id = 152 2025-01-08T02:22:29,918 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=152, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=152, snapshot={ ss=snapshot-testExportExpiredSnapshot table=testExportExpiredSnapshot type=FLUSH ttl=10 } in 428 msec 2025-01-08T02:22:30,096 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=152 2025-01-08T02:22:30,096 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testExportExpiredSnapshot, procId: 152 completed 2025-01-08T02:22:30,397 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0006_000001 (auth:SIMPLE) from 127.0.0.1:39614 2025-01-08T02:22:31,516 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:22:35,510 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0006/container_1736302816589_0006_01_000001/launch_container.sh] 2025-01-08T02:22:35,510 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0006/container_1736302816589_0006_01_000001/container_tokens] 2025-01-08T02:22:35,510 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0006/container_1736302816589_0006_01_000001/sysfs] 2025-01-08T02:22:36,390 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:22:38,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot 2025-01-08T02:22:38,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot Metrics about Tables on a single HBase RegionServer 2025-01-08T02:22:40,106 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302960106 2025-01-08T02:22:40,107 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302960106, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302960106, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:40,152 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:40,153 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302960106, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302960106/.hbase-snapshot/.tmp/snapshot-testExportExpiredSnapshot 2025-01-08T02:22:40,155 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:22:40,156 ERROR [Time-limited test {}] util.AbstractHBaseTool(153): Error running command-line tool org.apache.hadoop.hbase.snapshot.SnapshotTTLExpiredException: TTL for snapshot 'snapshot-testExportExpiredSnapshot' has already expired. at org.apache.hadoop.hbase.snapshot.ExportSnapshot.verifySnapshot(ExportSnapshot.java:948) ~[classes/:?] at org.apache.hadoop.hbase.snapshot.ExportSnapshot.doWork(ExportSnapshot.java:1093) ~[classes/:?] at org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:151) ~[hbase-common-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:82) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.runExportSnapshot(TestExportSnapshot.java:523) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportExpiredSnapshot(TestExportSnapshot.java:315) ~[test-classes/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) ~[junit-4.13.2.jar:4.13.2] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:22:40,158 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,158 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,159 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=155, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,161 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T02:22:40,162 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302960161"}]},"ts":"1736302960161"} 2025-01-08T02:22:40,163 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=DISABLING in hbase:meta 2025-01-08T02:22:40,165 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportExpiredSnapshot to state=DISABLING 2025-01-08T02:22:40,166 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=156, ppid=155, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportExpiredSnapshot}] 2025-01-08T02:22:40,167 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=157, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=f80872b7eb7ebb6d6d7c87636cc1f232, UNASSIGN}, {pid=158, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8235ae7ab56bd206f032e5c60426552e, UNASSIGN}] 2025-01-08T02:22:40,168 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=158, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8235ae7ab56bd206f032e5c60426552e, UNASSIGN 2025-01-08T02:22:40,168 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=157, ppid=156, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=f80872b7eb7ebb6d6d7c87636cc1f232, UNASSIGN 2025-01-08T02:22:40,169 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=158 updating hbase:meta row=8235ae7ab56bd206f032e5c60426552e, regionState=CLOSING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:40,169 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=157 updating hbase:meta row=f80872b7eb7ebb6d6d7c87636cc1f232, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:40,170 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:40,170 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=159, ppid=157, state=RUNNABLE; CloseRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:40,171 DEBUG [PEWorker-2 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:40,171 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=160, ppid=158, state=RUNNABLE; CloseRegionProcedure 8235ae7ab56bd206f032e5c60426552e, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:22:40,262 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T02:22:40,321 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:40,322 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(124): Close f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:40,322 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:40,322 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1681): Closing f80872b7eb7ebb6d6d7c87636cc1f232, disabling compactions & flushes 2025-01-08T02:22:40,322 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:40,322 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:40,322 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. after waiting 0 ms 2025-01-08T02:22:40,322 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:40,323 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:22:40,323 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(124): Close 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:40,324 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:40,324 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1681): Closing 8235ae7ab56bd206f032e5c60426552e, disabling compactions & flushes 2025-01-08T02:22:40,324 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1703): Closing region testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:40,324 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:40,324 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. after waiting 0 ms 2025-01-08T02:22:40,324 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:40,341 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:40,342 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:40,342 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232. 2025-01-08T02:22:40,342 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] regionserver.HRegion(1635): Region close journal for f80872b7eb7ebb6d6d7c87636cc1f232: 2025-01-08T02:22:40,344 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=159}] handler.UnassignRegionHandler(170): Closed f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:40,344 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=157 updating hbase:meta row=f80872b7eb7ebb6d6d7c87636cc1f232, regionState=CLOSED 2025-01-08T02:22:40,345 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:40,346 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:40,346 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1922): Closed testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e. 2025-01-08T02:22:40,346 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] regionserver.HRegion(1635): Region close journal for 8235ae7ab56bd206f032e5c60426552e: 2025-01-08T02:22:40,347 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=160}] handler.UnassignRegionHandler(170): Closed 8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:40,347 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=158 updating hbase:meta row=8235ae7ab56bd206f032e5c60426552e, regionState=CLOSED 2025-01-08T02:22:40,348 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=159, resume processing ppid=157 2025-01-08T02:22:40,348 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=159, ppid=157, state=SUCCESS; CloseRegionProcedure f80872b7eb7ebb6d6d7c87636cc1f232, server=c4e5d52b9759,39911,1736302808913 in 176 msec 2025-01-08T02:22:40,351 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=157, ppid=156, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=f80872b7eb7ebb6d6d7c87636cc1f232, UNASSIGN in 181 msec 2025-01-08T02:22:40,353 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=160, resume processing ppid=158 2025-01-08T02:22:40,353 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=160, ppid=158, state=SUCCESS; CloseRegionProcedure 8235ae7ab56bd206f032e5c60426552e, server=c4e5d52b9759,36743,1736302808837 in 178 msec 2025-01-08T02:22:40,359 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=158, resume processing ppid=156 2025-01-08T02:22:40,359 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=158, ppid=156, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportExpiredSnapshot, region=8235ae7ab56bd206f032e5c60426552e, UNASSIGN in 187 msec 2025-01-08T02:22:40,362 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=156, resume processing ppid=155 2025-01-08T02:22:40,362 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=156, ppid=155, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportExpiredSnapshot in 194 msec 2025-01-08T02:22:40,364 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302960363"}]},"ts":"1736302960363"} 2025-01-08T02:22:40,365 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportExpiredSnapshot, state=DISABLED in hbase:meta 2025-01-08T02:22:40,369 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(296): Set testtb-testExportExpiredSnapshot to state=DISABLED 2025-01-08T02:22:40,372 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=155, state=SUCCESS; DisableTableProcedure table=testtb-testExportExpiredSnapshot in 212 msec 2025-01-08T02:22:40,464 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=155 2025-01-08T02:22:40,464 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 155 completed 2025-01-08T02:22:40,465 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,466 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=161, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,467 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=161, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,468 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=161, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,468 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,473 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,477 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,477 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T02:22:40,477 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T02:22:40,477 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T02:22:40,477 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportExpiredSnapshot with data PBUF 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:40,479 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:40,480 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=161 2025-01-08T02:22:40,480 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:40,480 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:40,480 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:40,480 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:40,491 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:40,493 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/recovered.edits] 2025-01-08T02:22:40,498 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/cf/ef109958ed694211b2962ffb83afeb47 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/cf/ef109958ed694211b2962ffb83afeb47 2025-01-08T02:22:40,501 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232/recovered.edits/9.seqid 2025-01-08T02:22:40,502 DEBUG [HFileArchiver-16 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:40,504 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:40,506 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/recovered.edits] 2025-01-08T02:22:40,510 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/cf/fcf21c0566d44aa78ae187a4bc31e4b3 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/cf/fcf21c0566d44aa78ae187a4bc31e4b3 2025-01-08T02:22:40,513 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e/recovered.edits/9.seqid 2025-01-08T02:22:40,514 DEBUG [HFileArchiver-17 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportExpiredSnapshot/8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:40,514 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportExpiredSnapshot regions 2025-01-08T02:22:40,514 DEBUG [PEWorker-1 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f 2025-01-08T02:22:40,515 DEBUG [PEWorker-1 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf] 2025-01-08T02:22:40,519 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b2025010877725646de9f49228a1405aab0279f12_8235ae7ab56bd206f032e5c60426552e to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/c4ca4238a0b923820dcc509a6f75849b2025010877725646de9f49228a1405aab0279f12_8235ae7ab56bd206f032e5c60426552e 2025-01-08T02:22:40,520 DEBUG [PEWorker-1 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e20250108f6f240ee10b84b798e43d1ad97e9a8bb_f80872b7eb7ebb6d6d7c87636cc1f232 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f/cf/d41d8cd98f00b204e9800998ecf8427e20250108f6f240ee10b84b798e43d1ad97e9a8bb_f80872b7eb7ebb6d6d7c87636cc1f232 2025-01-08T02:22:40,521 DEBUG [PEWorker-1 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportExpiredSnapshot/3d21f28acea2939462cb18e5d9576f2f 2025-01-08T02:22:40,523 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=161, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,531 WARN [PEWorker-1 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportExpiredSnapshot from hbase:meta 2025-01-08T02:22:40,535 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportExpiredSnapshot' descriptor. 2025-01-08T02:22:40,541 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=161, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,541 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportExpiredSnapshot' from region states. 2025-01-08T02:22:40,541 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302960541"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:40,541 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302960541"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:40,558 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:22:40,558 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => f80872b7eb7ebb6d6d7c87636cc1f232, NAME => 'testtb-testExportExpiredSnapshot,,1736302946700.f80872b7eb7ebb6d6d7c87636cc1f232.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 8235ae7ab56bd206f032e5c60426552e, NAME => 'testtb-testExportExpiredSnapshot,1,1736302946700.8235ae7ab56bd206f032e5c60426552e.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:22:40,558 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportExpiredSnapshot' as deleted. 2025-01-08T02:22:40,559 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportExpiredSnapshot","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302960559"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:40,562 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportExpiredSnapshot state from META 2025-01-08T02:22:40,565 DEBUG [PEWorker-1 {}] procedure.DeleteTableProcedure(133): Finished pid=161, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportExpiredSnapshot 2025-01-08T02:22:40,566 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=161, state=SUCCESS; DeleteTableProcedure table=testtb-testExportExpiredSnapshot in 100 msec 2025-01-08T02:22:40,581 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=161 2025-01-08T02:22:40,581 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportExpiredSnapshot, procId: 161 completed 2025-01-08T02:22:40,606 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportExpiredSnapshot" 2025-01-08T02:22:40,608 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportExpiredSnapshot 2025-01-08T02:22:40,610 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snapshot-testExportExpiredSnapshot" 2025-01-08T02:22:40,611 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snapshot-testExportExpiredSnapshot 2025-01-08T02:22:40,613 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportExpiredSnapshot" 2025-01-08T02:22:40,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportExpiredSnapshot 2025-01-08T02:22:40,646 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportExpiredSnapshot Thread=792 (was 801), OpenFileDescriptor=799 (was 809), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=565 (was 668), ProcessCount=14 (was 17), AvailableMemoryMB=2515 (was 2183) - AvailableMemoryMB LEAK? - 2025-01-08T02:22:40,646 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=792 is superior to 500 2025-01-08T02:22:40,669 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testEmptyExportFileSystemState Thread=792, OpenFileDescriptor=799, MaxFileDescriptor=1048576, SystemLoadAverage=565, ProcessCount=14, AvailableMemoryMB=2513 2025-01-08T02:22:40,670 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=792 is superior to 500 2025-01-08T02:22:40,671 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:22:40,672 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=162, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:40,673 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:22:40,674 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testEmptyExportFileSystemState" procId is: 162 2025-01-08T02:22:40,674 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:22:40,675 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T02:22:40,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742222_1398 (size=448) 2025-01-08T02:22:40,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742222_1398 (size=448) 2025-01-08T02:22:40,688 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742222_1398 (size=448) 2025-01-08T02:22:40,692 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => abc47aa081596c09e48ebf249fd1a2b2, NAME => 'testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:40,692 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 8e7bf6158c7b7488f7c41a1178e903cd, NAME => 'testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testEmptyExportFileSystemState', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:40,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742224_1400 (size=73) 2025-01-08T02:22:40,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742224_1400 (size=73) 2025-01-08T02:22:40,714 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742223_1399 (size=73) 2025-01-08T02:22:40,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742223_1399 (size=73) 2025-01-08T02:22:40,715 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742223_1399 (size=73) 2025-01-08T02:22:40,716 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742224_1400 (size=73) 2025-01-08T02:22:40,717 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:40,717 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1681): Closing abc47aa081596c09e48ebf249fd1a2b2, disabling compactions & flushes 2025-01-08T02:22:40,717 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:40,717 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:40,717 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. after waiting 0 ms 2025-01-08T02:22:40,717 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:40,717 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:40,717 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-0 {}] regionserver.HRegion(1635): Region close journal for abc47aa081596c09e48ebf249fd1a2b2: 2025-01-08T02:22:40,721 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:40,721 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1681): Closing 8e7bf6158c7b7488f7c41a1178e903cd, disabling compactions & flushes 2025-01-08T02:22:40,721 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:40,721 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:40,721 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. after waiting 0 ms 2025-01-08T02:22:40,721 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:40,721 INFO [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:40,721 DEBUG [RegionOpenAndInit-testtb-testEmptyExportFileSystemState-pool-1 {}] regionserver.HRegion(1635): Region close journal for 8e7bf6158c7b7488f7c41a1178e903cd: 2025-01-08T02:22:40,723 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:22:40,724 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.","families":{"info":[{"qualifier":"regioninfo","vlen":72,"tag":[],"timestamp":"1736302960723"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302960723"}]},"ts":"1736302960723"} 2025-01-08T02:22:40,724 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.","families":{"info":[{"qualifier":"regioninfo","vlen":72,"tag":[],"timestamp":"1736302960723"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302960723"}]},"ts":"1736302960723"} 2025-01-08T02:22:40,726 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:22:40,727 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:22:40,727 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302960727"}]},"ts":"1736302960727"} 2025-01-08T02:22:40,729 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=ENABLING in hbase:meta 2025-01-08T02:22:40,733 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:22:40,734 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:22:40,734 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:22:40,734 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:22:40,734 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:22:40,734 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:22:40,734 INFO [PEWorker-5 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:22:40,734 DEBUG [PEWorker-5 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:22:40,734 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=abc47aa081596c09e48ebf249fd1a2b2, ASSIGN}, {pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=8e7bf6158c7b7488f7c41a1178e903cd, ASSIGN}] 2025-01-08T02:22:40,736 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=abc47aa081596c09e48ebf249fd1a2b2, ASSIGN 2025-01-08T02:22:40,736 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=8e7bf6158c7b7488f7c41a1178e903cd, ASSIGN 2025-01-08T02:22:40,737 INFO [PEWorker-4 {}] assignment.TransitRegionStateProcedure(264): Starting pid=163, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=abc47aa081596c09e48ebf249fd1a2b2, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:22:40,738 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=164, ppid=162, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=8e7bf6158c7b7488f7c41a1178e903cd, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:22:40,776 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T02:22:40,888 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:22:40,888 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=163 updating hbase:meta row=abc47aa081596c09e48ebf249fd1a2b2, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:40,888 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=164 updating hbase:meta row=8e7bf6158c7b7488f7c41a1178e903cd, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:40,891 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=165, ppid=164, state=RUNNABLE; OpenRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:40,893 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=166, ppid=163, state=RUNNABLE; OpenRegionProcedure abc47aa081596c09e48ebf249fd1a2b2, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:40,977 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T02:22:41,044 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:41,046 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:41,048 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] handler.AssignRegionHandler(135): Open testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:41,048 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7285): Opening region: {ENCODED => 8e7bf6158c7b7488f7c41a1178e903cd, NAME => 'testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:22:41,048 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. service=AccessControlService 2025-01-08T02:22:41,048 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:41,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testEmptyExportFileSystemState 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:41,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7327): checking encryption for 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,049 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(7330): checking classloading for 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,050 INFO [StoreOpener-8e7bf6158c7b7488f7c41a1178e903cd-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,051 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] handler.AssignRegionHandler(135): Open testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,051 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7285): Opening region: {ENCODED => abc47aa081596c09e48ebf249fd1a2b2, NAME => 'testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:22:41,052 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. service=AccessControlService 2025-01-08T02:22:41,052 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:41,052 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testEmptyExportFileSystemState abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,052 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(894): Instantiated testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:41,052 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7327): checking encryption for abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,052 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(7330): checking classloading for abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,052 INFO [StoreOpener-8e7bf6158c7b7488f7c41a1178e903cd-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 8e7bf6158c7b7488f7c41a1178e903cd columnFamilyName cf 2025-01-08T02:22:41,054 DEBUG [StoreOpener-8e7bf6158c7b7488f7c41a1178e903cd-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:41,054 INFO [StoreOpener-abc47aa081596c09e48ebf249fd1a2b2-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,054 INFO [StoreOpener-8e7bf6158c7b7488f7c41a1178e903cd-1 {}] regionserver.HStore(327): Store=8e7bf6158c7b7488f7c41a1178e903cd/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:41,055 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,055 INFO [StoreOpener-abc47aa081596c09e48ebf249fd1a2b2-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region abc47aa081596c09e48ebf249fd1a2b2 columnFamilyName cf 2025-01-08T02:22:41,056 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,057 DEBUG [StoreOpener-abc47aa081596c09e48ebf249fd1a2b2-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:41,057 INFO [StoreOpener-abc47aa081596c09e48ebf249fd1a2b2-1 {}] regionserver.HStore(327): Store=abc47aa081596c09e48ebf249fd1a2b2/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:41,059 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,060 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,061 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1085): writing seq id for 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,062 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1085): writing seq id for abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,063 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:41,064 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1102): Opened 8e7bf6158c7b7488f7c41a1178e903cd; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=63387150, jitterRate=-0.055457860231399536}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:41,065 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegion(1001): Region open journal for 8e7bf6158c7b7488f7c41a1178e903cd: 2025-01-08T02:22:41,065 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd., pid=165, masterSystemTime=1736302961044 2025-01-08T02:22:41,067 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:41,067 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=165}] handler.AssignRegionHandler(164): Opened testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:41,068 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:41,068 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=164 updating hbase:meta row=8e7bf6158c7b7488f7c41a1178e903cd, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:41,068 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1102): Opened abc47aa081596c09e48ebf249fd1a2b2; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=61940308, jitterRate=-0.07701748609542847}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:41,068 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegion(1001): Region open journal for abc47aa081596c09e48ebf249fd1a2b2: 2025-01-08T02:22:41,070 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2., pid=166, masterSystemTime=1736302961046 2025-01-08T02:22:41,073 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,073 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=166}] handler.AssignRegionHandler(164): Opened testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,074 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=163 updating hbase:meta row=abc47aa081596c09e48ebf249fd1a2b2, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:41,075 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=165, resume processing ppid=164 2025-01-08T02:22:41,076 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=165, ppid=164, state=SUCCESS; OpenRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd, server=c4e5d52b9759,37031,1736302808764 in 181 msec 2025-01-08T02:22:41,077 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=164, ppid=162, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=8e7bf6158c7b7488f7c41a1178e903cd, ASSIGN in 342 msec 2025-01-08T02:22:41,078 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=166, resume processing ppid=163 2025-01-08T02:22:41,078 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=166, ppid=163, state=SUCCESS; OpenRegionProcedure abc47aa081596c09e48ebf249fd1a2b2, server=c4e5d52b9759,39911,1736302808913 in 184 msec 2025-01-08T02:22:41,080 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=163, resume processing ppid=162 2025-01-08T02:22:41,081 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=163, ppid=162, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=abc47aa081596c09e48ebf249fd1a2b2, ASSIGN in 344 msec 2025-01-08T02:22:41,081 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:22:41,081 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302961081"}]},"ts":"1736302961081"} 2025-01-08T02:22:41,082 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=ENABLED in hbase:meta 2025-01-08T02:22:41,085 INFO [PEWorker-5 {}] procedure.CreateTableProcedure(89): pid=162, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testEmptyExportFileSystemState execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:22:41,085 DEBUG [PEWorker-5 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testEmptyExportFileSystemState jenkins: RWXCA 2025-01-08T02:22:41,087 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T02:22:41,094 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:41,094 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:41,094 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:41,094 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:41,097 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,097 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,097 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,098 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,098 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,098 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,099 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF\x0AK\x0A\x07jenkins\x12@\x08\x03"<\x0A0\x0A\x07default\x12%testtb-testEmptyExportFileSystemState \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,099 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:41,099 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=162, state=SUCCESS; CreateTableProcedure table=testtb-testEmptyExportFileSystemState in 426 msec 2025-01-08T02:22:41,279 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=162 2025-01-08T02:22:41,279 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 162 completed 2025-01-08T02:22:41,280 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testEmptyExportFileSystemState' 2025-01-08T02:22:41,282 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testEmptyExportFileSystemState 2025-01-08T02:22:41,282 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,283 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:41,305 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:22:41,306 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302961306 (current time:1736302961306). 2025-01-08T02:22:41,306 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:41,306 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testEmptyExportFileSystemState VERSION not specified, setting to 2 2025-01-08T02:22:41,306 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:41,307 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x16e83902 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6ede0e34 2025-01-08T02:22:41,312 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3824e48e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:41,314 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:41,315 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41746, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:41,316 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x16e83902 to 127.0.0.1:61593 2025-01-08T02:22:41,316 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:41,317 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x73a9b1d7 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@6d18be38 2025-01-08T02:22:41,332 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@72872781, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:41,337 DEBUG [hconnection-0x23eb10b2-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:41,338 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41748, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:41,339 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:41,340 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46534, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:41,341 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x73a9b1d7 to 127.0.0.1:61593 2025-01-08T02:22:41,341 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:41,341 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T02:22:41,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:41,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=167, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:22:41,343 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 167 2025-01-08T02:22:41,344 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:41,344 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T02:22:41,345 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:41,347 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:41,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742225_1401 (size=185) 2025-01-08T02:22:41,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742225_1401 (size=185) 2025-01-08T02:22:41,377 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742225_1401 (size=185) 2025-01-08T02:22:41,379 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:41,379 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2}, {pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd}] 2025-01-08T02:22:41,380 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,380 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,445 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T02:22:41,535 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:41,536 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=168 2025-01-08T02:22:41,536 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:41,536 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,537 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.HRegion(2538): Flush status journal for abc47aa081596c09e48ebf249fd1a2b2: 2025-01-08T02:22:41,537 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. for emptySnaptb0-testEmptyExportFileSystemState completed. 2025-01-08T02:22:41,537 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.' region-info for snapshot=emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,537 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:41,537 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:41,537 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=169 2025-01-08T02:22:41,538 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:41,538 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.HRegion(2538): Flush status journal for 8e7bf6158c7b7488f7c41a1178e903cd: 2025-01-08T02:22:41,538 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. for emptySnaptb0-testEmptyExportFileSystemState completed. 2025-01-08T02:22:41,539 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.' region-info for snapshot=emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,539 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:41,539 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:41,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742226_1402 (size=76) 2025-01-08T02:22:41,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742226_1402 (size=76) 2025-01-08T02:22:41,547 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742226_1402 (size=76) 2025-01-08T02:22:41,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,548 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=168}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=168 2025-01-08T02:22:41,548 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=168 2025-01-08T02:22:41,548 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testEmptyExportFileSystemState on region abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,549 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=168, ppid=167, state=RUNNABLE; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,552 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=168, ppid=167, state=SUCCESS; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2 in 170 msec 2025-01-08T02:22:41,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742227_1403 (size=76) 2025-01-08T02:22:41,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742227_1403 (size=76) 2025-01-08T02:22:41,559 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742227_1403 (size=76) 2025-01-08T02:22:41,560 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:41,560 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=169}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=169 2025-01-08T02:22:41,560 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=169 2025-01-08T02:22:41,560 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testEmptyExportFileSystemState on region 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,560 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=169, ppid=167, state=RUNNABLE; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,563 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:41,563 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=169, resume processing ppid=167 2025-01-08T02:22:41,564 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:41,564 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=169, ppid=167, state=SUCCESS; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd in 182 msec 2025-01-08T02:22:41,565 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:41,565 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:41,565 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:41,566 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:22:41,582 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742228_1404 (size=68) 2025-01-08T02:22:41,582 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742228_1404 (size=68) 2025-01-08T02:22:41,583 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742228_1404 (size=68) 2025-01-08T02:22:41,584 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:41,584 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,585 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742229_1405 (size=673) 2025-01-08T02:22:41,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742229_1405 (size=673) 2025-01-08T02:22:41,602 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742229_1405 (size=673) 2025-01-08T02:22:41,613 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:41,621 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:41,622 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,630 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=167, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:41,630 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 167 2025-01-08T02:22:41,631 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=167, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=167, snapshot={ ss=emptySnaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } in 287 msec 2025-01-08T02:22:41,649 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=167 2025-01-08T02:22:41,650 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 167 completed 2025-01-08T02:22:41,654 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.HRegion(8254): writing data to region testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:41,657 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:41,663 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testEmptyExportFileSystemState 2025-01-08T02:22:41,663 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,664 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:41,680 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:22:41,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302961680 (current time:1736302961680). 2025-01-08T02:22:41,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:41,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testEmptyExportFileSystemState VERSION not specified, setting to 2 2025-01-08T02:22:41,680 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:41,681 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x7ec33cf1 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@556cd1b 2025-01-08T02:22:41,685 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@9f8cd02, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:41,689 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:41,692 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41764, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:41,693 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x7ec33cf1 to 127.0.0.1:61593 2025-01-08T02:22:41,693 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:41,694 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x46576a45 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@75da4981 2025-01-08T02:22:41,707 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3a851ae5, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:41,713 DEBUG [hconnection-0x3c6686f9-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:41,714 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:41772, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:41,715 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:41,716 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:46540, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:41,717 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x46576a45 to 127.0.0.1:61593 2025-01-08T02:22:41,717 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:41,717 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testEmptyExportFileSystemState], kv [jenkins: RWXCA] 2025-01-08T02:22:41,718 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:41,719 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=170, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } 2025-01-08T02:22:41,719 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 170 2025-01-08T02:22:41,720 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:41,721 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T02:22:41,721 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:41,723 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:41,736 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742230_1406 (size=180) 2025-01-08T02:22:41,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742230_1406 (size=180) 2025-01-08T02:22:41,737 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742230_1406 (size=180) 2025-01-08T02:22:41,739 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:41,739 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2}, {pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd}] 2025-01-08T02:22:41,740 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,740 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,822 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T02:22:41,891 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:41,891 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:41,892 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=171 2025-01-08T02:22:41,892 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=172 2025-01-08T02:22:41,892 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:41,892 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,893 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(2837): Flushing 8e7bf6158c7b7488f7c41a1178e903cd 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T02:22:41,893 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(2837): Flushing abc47aa081596c09e48ebf249fd1a2b2 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T02:22:41,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108913052a3509c400fb0d2e5f57b348ff8_abc47aa081596c09e48ebf249fd1a2b2 is 71, key is 07779b290e447d27a5e18d52e3b789bd/cf:q/1736302961657/Put/seqid=0 2025-01-08T02:22:41,918 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010839d7488d270d44ac867c182ba7b0970d_8e7bf6158c7b7488f7c41a1178e903cd is 71, key is 1e0cda7d929262ca5f20e217c5dffe6a/cf:q/1736302961654/Put/seqid=0 2025-01-08T02:22:41,929 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742232_1408 (size=8102) 2025-01-08T02:22:41,929 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742232_1408 (size=8102) 2025-01-08T02:22:41,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742232_1408 (size=8102) 2025-01-08T02:22:41,930 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742231_1407 (size=5171) 2025-01-08T02:22:41,930 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:41,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742231_1407 (size=5171) 2025-01-08T02:22:41,931 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742231_1407 (size=5171) 2025-01-08T02:22:41,931 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:41,936 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010839d7488d270d44ac867c182ba7b0970d_8e7bf6158c7b7488f7c41a1178e903cd to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010839d7488d270d44ac867c182ba7b0970d_8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,937 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/.tmp/cf/24c0473f078d4e568aacc70d1ab387da, store: [table=testtb-testEmptyExportFileSystemState family=cf region=8e7bf6158c7b7488f7c41a1178e903cd] 2025-01-08T02:22:41,937 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108913052a3509c400fb0d2e5f57b348ff8_abc47aa081596c09e48ebf249fd1a2b2 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108913052a3509c400fb0d2e5f57b348ff8_abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,938 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/.tmp/cf/24c0473f078d4e568aacc70d1ab387da is 214, key is 169dc5190a2878a4c0c707fdd38d13e8d/cf:q/1736302961654/Put/seqid=0 2025-01-08T02:22:41,938 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/.tmp/cf/b7f75b25fc2d436785fdccef2081aef0, store: [table=testtb-testEmptyExportFileSystemState family=cf region=abc47aa081596c09e48ebf249fd1a2b2] 2025-01-08T02:22:41,939 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/.tmp/cf/b7f75b25fc2d436785fdccef2081aef0 is 214, key is 07a388a67e9ba9308954e2a49c4ebc5a0/cf:q/1736302961657/Put/seqid=0 2025-01-08T02:22:41,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742234_1410 (size=6146) 2025-01-08T02:22:41,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742234_1410 (size=6146) 2025-01-08T02:22:41,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742234_1410 (size=6146) 2025-01-08T02:22:41,952 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/.tmp/cf/b7f75b25fc2d436785fdccef2081aef0 2025-01-08T02:22:41,956 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/.tmp/cf/b7f75b25fc2d436785fdccef2081aef0 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/cf/b7f75b25fc2d436785fdccef2081aef0 2025-01-08T02:22:41,960 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/cf/b7f75b25fc2d436785fdccef2081aef0, entries=4, sequenceid=6, filesize=6.0 K 2025-01-08T02:22:41,961 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for abc47aa081596c09e48ebf249fd1a2b2 in 68ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:41,961 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.HRegion(2538): Flush status journal for abc47aa081596c09e48ebf249fd1a2b2: 2025-01-08T02:22:41,961 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. for snaptb0-testEmptyExportFileSystemState completed. 2025-01-08T02:22:41,961 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.' region-info for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,961 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:41,961 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/cf/b7f75b25fc2d436785fdccef2081aef0] hfiles 2025-01-08T02:22:41,961 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/cf/b7f75b25fc2d436785fdccef2081aef0 for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742233_1409 (size=15029) 2025-01-08T02:22:41,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742233_1409 (size=15029) 2025-01-08T02:22:41,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742233_1409 (size=15029) 2025-01-08T02:22:41,967 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/.tmp/cf/24c0473f078d4e568aacc70d1ab387da 2025-01-08T02:22:41,980 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/.tmp/cf/24c0473f078d4e568aacc70d1ab387da as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/cf/24c0473f078d4e568aacc70d1ab387da 2025-01-08T02:22:41,985 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/cf/24c0473f078d4e568aacc70d1ab387da, entries=46, sequenceid=6, filesize=14.7 K 2025-01-08T02:22:41,987 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for 8e7bf6158c7b7488f7c41a1178e903cd in 94ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:41,987 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.HRegion(2538): Flush status journal for 8e7bf6158c7b7488f7c41a1178e903cd: 2025-01-08T02:22:41,987 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. for snaptb0-testEmptyExportFileSystemState completed. 2025-01-08T02:22:41,987 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742235_1411 (size=115) 2025-01-08T02:22:41,987 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(241): Storing 'testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.' region-info for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,987 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:41,987 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/cf/24c0473f078d4e568aacc70d1ab387da] hfiles 2025-01-08T02:22:41,987 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/cf/24c0473f078d4e568aacc70d1ab387da for snapshot=snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:41,988 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742235_1411 (size=115) 2025-01-08T02:22:41,989 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742235_1411 (size=115) 2025-01-08T02:22:41,989 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:41,989 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=171}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=171 2025-01-08T02:22:41,990 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=171 2025-01-08T02:22:41,990 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testEmptyExportFileSystemState on region abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,990 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=171, ppid=170, state=RUNNABLE; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:41,992 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=171, ppid=170, state=SUCCESS; SnapshotRegionProcedure abc47aa081596c09e48ebf249fd1a2b2 in 252 msec 2025-01-08T02:22:41,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742236_1412 (size=115) 2025-01-08T02:22:41,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742236_1412 (size=115) 2025-01-08T02:22:41,995 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742236_1412 (size=115) 2025-01-08T02:22:41,996 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:41,996 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=172}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=172 2025-01-08T02:22:41,996 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=172 2025-01-08T02:22:41,996 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testEmptyExportFileSystemState on region 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,996 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=172, ppid=170, state=RUNNABLE; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:41,998 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=172, resume processing ppid=170 2025-01-08T02:22:41,998 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=172, ppid=170, state=SUCCESS; SnapshotRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd in 258 msec 2025-01-08T02:22:41,998 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:41,999 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:41,999 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:42,000 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:42,000 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:42,001 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010839d7488d270d44ac867c182ba7b0970d_8e7bf6158c7b7488f7c41a1178e903cd, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108913052a3509c400fb0d2e5f57b348ff8_abc47aa081596c09e48ebf249fd1a2b2] hfiles 2025-01-08T02:22:42,001 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010839d7488d270d44ac867c182ba7b0970d_8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:42,001 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108913052a3509c400fb0d2e5f57b348ff8_abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:42,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742237_1413 (size=299) 2025-01-08T02:22:42,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742237_1413 (size=299) 2025-01-08T02:22:42,015 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742237_1413 (size=299) 2025-01-08T02:22:42,016 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:42,016 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:42,017 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:42,023 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T02:22:42,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742238_1414 (size=983) 2025-01-08T02:22:42,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742238_1414 (size=983) 2025-01-08T02:22:42,035 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742238_1414 (size=983) 2025-01-08T02:22:42,042 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:42,049 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:42,050 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testEmptyExportFileSystemState to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:42,051 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=170, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:42,051 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 }, snapshot procedure id = 170 2025-01-08T02:22:42,052 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=170, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=170, snapshot={ ss=snaptb0-testEmptyExportFileSystemState table=testtb-testEmptyExportFileSystemState type=FLUSH ttl=0 } in 333 msec 2025-01-08T02:22:42,324 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=170 2025-01-08T02:22:42,324 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 170 completed 2025-01-08T02:22:42,324 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324 2025-01-08T02:22:42,324 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:42,357 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:42,357 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:42,359 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:22:42,364 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324/.hbase-snapshot/.tmp/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:42,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742239_1415 (size=185) 2025-01-08T02:22:42,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742239_1415 (size=185) 2025-01-08T02:22:42,395 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742240_1416 (size=673) 2025-01-08T02:22:42,396 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742240_1416 (size=673) 2025-01-08T02:22:42,398 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742240_1416 (size=673) 2025-01-08T02:22:42,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742239_1415 (size=185) 2025-01-08T02:22:42,410 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:42,411 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:42,411 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:42,411 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,535 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-8148915952476275638.jar 2025-01-08T02:22:43,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,536 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,633 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-13986979867890324336.jar 2025-01-08T02:22:43,633 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,634 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,634 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,635 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,635 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,635 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:43,636 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:22:43,636 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:22:43,636 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:22:43,636 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:22:43,637 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:22:43,637 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:22:43,637 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:22:43,637 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:22:43,638 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:22:43,638 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:22:43,638 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:22:43,639 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:22:43,639 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:43,639 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:43,640 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:43,640 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:43,640 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:43,641 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:43,641 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:43,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742241_1417 (size=127628) 2025-01-08T02:22:43,711 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742241_1417 (size=127628) 2025-01-08T02:22:43,712 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742241_1417 (size=127628) 2025-01-08T02:22:43,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T02:22:43,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T02:22:43,735 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742242_1418 (size=2172137) 2025-01-08T02:22:43,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742243_1419 (size=213228) 2025-01-08T02:22:43,751 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742243_1419 (size=213228) 2025-01-08T02:22:43,752 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742243_1419 (size=213228) 2025-01-08T02:22:43,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T02:22:43,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T02:22:43,766 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742244_1420 (size=1877034) 2025-01-08T02:22:43,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742245_1421 (size=533455) 2025-01-08T02:22:43,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742245_1421 (size=533455) 2025-01-08T02:22:43,788 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742245_1421 (size=533455) 2025-01-08T02:22:43,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742246_1422 (size=7280644) 2025-01-08T02:22:43,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742246_1422 (size=7280644) 2025-01-08T02:22:43,830 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742246_1422 (size=7280644) 2025-01-08T02:22:43,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742247_1423 (size=4188619) 2025-01-08T02:22:43,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742247_1423 (size=4188619) 2025-01-08T02:22:43,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742247_1423 (size=4188619) 2025-01-08T02:22:43,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742248_1424 (size=20406) 2025-01-08T02:22:43,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742248_1424 (size=20406) 2025-01-08T02:22:43,887 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742248_1424 (size=20406) 2025-01-08T02:22:43,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742249_1425 (size=75495) 2025-01-08T02:22:43,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742249_1425 (size=75495) 2025-01-08T02:22:43,904 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742249_1425 (size=75495) 2025-01-08T02:22:43,921 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742250_1426 (size=45609) 2025-01-08T02:22:43,921 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742250_1426 (size=45609) 2025-01-08T02:22:43,921 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742250_1426 (size=45609) 2025-01-08T02:22:44,341 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742251_1427 (size=451756) 2025-01-08T02:22:44,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742251_1427 (size=451756) 2025-01-08T02:22:44,342 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742251_1427 (size=451756) 2025-01-08T02:22:44,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742252_1428 (size=110084) 2025-01-08T02:22:44,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742252_1428 (size=110084) 2025-01-08T02:22:44,372 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742252_1428 (size=110084) 2025-01-08T02:22:44,393 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742253_1429 (size=1323991) 2025-01-08T02:22:44,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742253_1429 (size=1323991) 2025-01-08T02:22:44,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742253_1429 (size=1323991) 2025-01-08T02:22:44,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742254_1430 (size=23076) 2025-01-08T02:22:44,414 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742254_1430 (size=23076) 2025-01-08T02:22:44,415 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742254_1430 (size=23076) 2025-01-08T02:22:44,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742255_1431 (size=126803) 2025-01-08T02:22:44,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742255_1431 (size=126803) 2025-01-08T02:22:44,431 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742255_1431 (size=126803) 2025-01-08T02:22:44,449 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742256_1432 (size=322274) 2025-01-08T02:22:44,450 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742256_1432 (size=322274) 2025-01-08T02:22:44,451 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742256_1432 (size=322274) 2025-01-08T02:22:44,493 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742257_1433 (size=6350715) 2025-01-08T02:22:44,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742257_1433 (size=6350715) 2025-01-08T02:22:44,494 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742257_1433 (size=6350715) 2025-01-08T02:22:44,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742258_1434 (size=1832290) 2025-01-08T02:22:44,512 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742258_1434 (size=1832290) 2025-01-08T02:22:44,513 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742258_1434 (size=1832290) 2025-01-08T02:22:44,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742259_1435 (size=30081) 2025-01-08T02:22:44,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742259_1435 (size=30081) 2025-01-08T02:22:44,525 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742259_1435 (size=30081) 2025-01-08T02:22:44,932 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742260_1436 (size=53616) 2025-01-08T02:22:44,933 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742260_1436 (size=53616) 2025-01-08T02:22:44,933 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742260_1436 (size=53616) 2025-01-08T02:22:44,939 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742261_1437 (size=29229) 2025-01-08T02:22:44,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742261_1437 (size=29229) 2025-01-08T02:22:44,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742261_1437 (size=29229) 2025-01-08T02:22:44,946 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742262_1438 (size=169089) 2025-01-08T02:22:44,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742262_1438 (size=169089) 2025-01-08T02:22:44,947 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742262_1438 (size=169089) 2025-01-08T02:22:44,968 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742263_1439 (size=5175431) 2025-01-08T02:22:44,969 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742263_1439 (size=5175431) 2025-01-08T02:22:44,969 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742263_1439 (size=5175431) 2025-01-08T02:22:44,976 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742264_1440 (size=136454) 2025-01-08T02:22:44,976 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742264_1440 (size=136454) 2025-01-08T02:22:44,976 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742264_1440 (size=136454) 2025-01-08T02:22:44,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742265_1441 (size=907469) 2025-01-08T02:22:44,984 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742265_1441 (size=907469) 2025-01-08T02:22:44,985 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742265_1441 (size=907469) 2025-01-08T02:22:44,999 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T02:22:44,999 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T02:22:44,999 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742266_1442 (size=3317408) 2025-01-08T02:22:45,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742267_1443 (size=503880) 2025-01-08T02:22:45,006 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742267_1443 (size=503880) 2025-01-08T02:22:45,007 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742267_1443 (size=503880) 2025-01-08T02:22:45,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T02:22:45,024 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T02:22:45,025 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742268_1444 (size=4695811) 2025-01-08T02:22:45,026 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:22:45,028 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'emptySnaptb0-testEmptyExportFileSystemState' hfile list 2025-01-08T02:22:45,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742269_1445 (size=7) 2025-01-08T02:22:45,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742269_1445 (size=7) 2025-01-08T02:22:45,034 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742269_1445 (size=7) 2025-01-08T02:22:45,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742270_1446 (size=10) 2025-01-08T02:22:45,039 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742270_1446 (size=10) 2025-01-08T02:22:45,040 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742270_1446 (size=10) 2025-01-08T02:22:45,051 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742271_1447 (size=304784) 2025-01-08T02:22:45,051 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742271_1447 (size=304784) 2025-01-08T02:22:45,052 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742271_1447 (size=304784) 2025-01-08T02:22:45,067 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:22:45,067 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:22:45,403 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0007_000001 (auth:SIMPLE) from 127.0.0.1:45260 2025-01-08T02:22:45,882 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:22:48,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState 2025-01-08T02:22:48,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState Metrics about Tables on a single HBase RegionServer 2025-01-08T02:22:48,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportExpiredSnapshot 2025-01-08T02:22:51,468 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0007_000001 (auth:SIMPLE) from 127.0.0.1:52312 2025-01-08T02:22:51,732 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742272_1448 (size=350434) 2025-01-08T02:22:51,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742272_1448 (size=350434) 2025-01-08T02:22:51,733 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742272_1448 (size=350434) 2025-01-08T02:22:52,567 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742273_1449 (size=8568) 2025-01-08T02:22:52,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742273_1449 (size=8568) 2025-01-08T02:22:52,568 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742273_1449 (size=8568) 2025-01-08T02:22:52,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742274_1450 (size=460) 2025-01-08T02:22:52,579 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742274_1450 (size=460) 2025-01-08T02:22:52,580 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742274_1450 (size=460) 2025-01-08T02:22:52,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742275_1451 (size=8568) 2025-01-08T02:22:52,614 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742275_1451 (size=8568) 2025-01-08T02:22:52,615 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742275_1451 (size=8568) 2025-01-08T02:22:52,645 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742276_1452 (size=350434) 2025-01-08T02:22:52,645 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742276_1452 (size=350434) 2025-01-08T02:22:52,645 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742276_1452 (size=350434) 2025-01-08T02:22:53,934 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:22:54,175 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:22:54,175 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:22:54,185 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:54,185 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:22:54,185 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:22:54,185 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:54,186 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/.snapshotinfo 2025-01-08T02:22:54,186 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/data.manifest 2025-01-08T02:22:54,186 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:54,187 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/.snapshotinfo 2025-01-08T02:22:54,187 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736302962324/.hbase-snapshot/emptySnaptb0-testEmptyExportFileSystemState/data.manifest 2025-01-08T02:22:54,191 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,191 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,192 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=173, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,194 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T02:22:54,194 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302974194"}]},"ts":"1736302974194"} 2025-01-08T02:22:54,196 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=DISABLING in hbase:meta 2025-01-08T02:22:54,198 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(284): Set testtb-testEmptyExportFileSystemState to state=DISABLING 2025-01-08T02:22:54,199 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=174, ppid=173, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testEmptyExportFileSystemState}] 2025-01-08T02:22:54,200 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=175, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=abc47aa081596c09e48ebf249fd1a2b2, UNASSIGN}, {pid=176, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=8e7bf6158c7b7488f7c41a1178e903cd, UNASSIGN}] 2025-01-08T02:22:54,201 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=176, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=8e7bf6158c7b7488f7c41a1178e903cd, UNASSIGN 2025-01-08T02:22:54,201 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=175, ppid=174, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=abc47aa081596c09e48ebf249fd1a2b2, UNASSIGN 2025-01-08T02:22:54,202 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=176 updating hbase:meta row=8e7bf6158c7b7488f7c41a1178e903cd, regionState=CLOSING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:54,202 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=175 updating hbase:meta row=abc47aa081596c09e48ebf249fd1a2b2, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:54,203 DEBUG [PEWorker-5 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:54,204 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=177, ppid=176, state=RUNNABLE; CloseRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:54,204 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:22:54,204 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=178, ppid=175, state=RUNNABLE; CloseRegionProcedure abc47aa081596c09e48ebf249fd1a2b2, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:54,295 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T02:22:54,355 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:54,356 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(124): Close 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:54,356 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:54,356 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1681): Closing 8e7bf6158c7b7488f7c41a1178e903cd, disabling compactions & flushes 2025-01-08T02:22:54,356 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:54,356 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:54,356 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. after waiting 0 ms 2025-01-08T02:22:54,356 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:54,356 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:54,357 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(124): Close abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:54,357 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:22:54,357 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1681): Closing abc47aa081596c09e48ebf249fd1a2b2, disabling compactions & flushes 2025-01-08T02:22:54,357 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1703): Closing region testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:54,357 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:54,357 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1791): Acquired close lock on testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. after waiting 0 ms 2025-01-08T02:22:54,357 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1801): Updates disabled for region testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:54,361 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:54,361 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:22:54,362 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:54,362 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:22:54,362 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd. 2025-01-08T02:22:54,362 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1922): Closed testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2. 2025-01-08T02:22:54,362 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] regionserver.HRegion(1635): Region close journal for abc47aa081596c09e48ebf249fd1a2b2: 2025-01-08T02:22:54,362 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] regionserver.HRegion(1635): Region close journal for 8e7bf6158c7b7488f7c41a1178e903cd: 2025-01-08T02:22:54,363 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=178}] handler.UnassignRegionHandler(170): Closed abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:54,364 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=175 updating hbase:meta row=abc47aa081596c09e48ebf249fd1a2b2, regionState=CLOSED 2025-01-08T02:22:54,364 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=177}] handler.UnassignRegionHandler(170): Closed 8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:54,365 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=176 updating hbase:meta row=8e7bf6158c7b7488f7c41a1178e903cd, regionState=CLOSED 2025-01-08T02:22:54,369 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=178, resume processing ppid=175 2025-01-08T02:22:54,369 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=178, ppid=175, state=SUCCESS; CloseRegionProcedure abc47aa081596c09e48ebf249fd1a2b2, server=c4e5d52b9759,39911,1736302808913 in 163 msec 2025-01-08T02:22:54,370 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=177, resume processing ppid=176 2025-01-08T02:22:54,370 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=175, ppid=174, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=abc47aa081596c09e48ebf249fd1a2b2, UNASSIGN in 169 msec 2025-01-08T02:22:54,370 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=177, ppid=176, state=SUCCESS; CloseRegionProcedure 8e7bf6158c7b7488f7c41a1178e903cd, server=c4e5d52b9759,37031,1736302808764 in 164 msec 2025-01-08T02:22:54,371 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=176, resume processing ppid=174 2025-01-08T02:22:54,371 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=176, ppid=174, state=SUCCESS; TransitRegionStateProcedure table=testtb-testEmptyExportFileSystemState, region=8e7bf6158c7b7488f7c41a1178e903cd, UNASSIGN in 170 msec 2025-01-08T02:22:54,373 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=174, resume processing ppid=173 2025-01-08T02:22:54,373 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=174, ppid=173, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testEmptyExportFileSystemState in 172 msec 2025-01-08T02:22:54,374 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302974373"}]},"ts":"1736302974373"} 2025-01-08T02:22:54,375 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testEmptyExportFileSystemState, state=DISABLED in hbase:meta 2025-01-08T02:22:54,376 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(296): Set testtb-testEmptyExportFileSystemState to state=DISABLED 2025-01-08T02:22:54,378 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=173, state=SUCCESS; DisableTableProcedure table=testtb-testEmptyExportFileSystemState in 185 msec 2025-01-08T02:22:54,496 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=173 2025-01-08T02:22:54,496 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 173 completed 2025-01-08T02:22:54,496 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,497 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=179, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,498 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=179, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,498 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,498 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=179, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,500 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,501 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:54,501 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:54,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,503 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,503 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/recovered.edits] 2025-01-08T02:22:54,503 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/recovered.edits] 2025-01-08T02:22:54,503 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T02:22:54,504 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T02:22:54,504 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T02:22:54,504 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testEmptyExportFileSystemState with data PBUF 2025-01-08T02:22:54,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,504 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,505 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,505 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,505 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:54,505 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:54,505 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:54,505 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:54,505 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=179 2025-01-08T02:22:54,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:54,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:54,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:54,506 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:54,507 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/cf/24c0473f078d4e568aacc70d1ab387da to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/cf/24c0473f078d4e568aacc70d1ab387da 2025-01-08T02:22:54,508 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/cf/b7f75b25fc2d436785fdccef2081aef0 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/cf/b7f75b25fc2d436785fdccef2081aef0 2025-01-08T02:22:54,510 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd/recovered.edits/9.seqid 2025-01-08T02:22:54,510 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2/recovered.edits/9.seqid 2025-01-08T02:22:54,511 DEBUG [HFileArchiver-12 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:54,511 DEBUG [HFileArchiver-18 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testEmptyExportFileSystemState/abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:54,511 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(313): Archived testtb-testEmptyExportFileSystemState regions 2025-01-08T02:22:54,511 DEBUG [PEWorker-4 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0 2025-01-08T02:22:54,512 DEBUG [PEWorker-4 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf] 2025-01-08T02:22:54,515 DEBUG [PEWorker-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010839d7488d270d44ac867c182ba7b0970d_8e7bf6158c7b7488f7c41a1178e903cd to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/c4ca4238a0b923820dcc509a6f75849b2025010839d7488d270d44ac867c182ba7b0970d_8e7bf6158c7b7488f7c41a1178e903cd 2025-01-08T02:22:54,516 DEBUG [PEWorker-4 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108913052a3509c400fb0d2e5f57b348ff8_abc47aa081596c09e48ebf249fd1a2b2 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0/cf/d41d8cd98f00b204e9800998ecf8427e20250108913052a3509c400fb0d2e5f57b348ff8_abc47aa081596c09e48ebf249fd1a2b2 2025-01-08T02:22:54,516 DEBUG [PEWorker-4 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testEmptyExportFileSystemState/37e3865d2ff7e5e2da33a03e3f723df0 2025-01-08T02:22:54,518 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=179, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,520 WARN [PEWorker-4 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testEmptyExportFileSystemState from hbase:meta 2025-01-08T02:22:54,522 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testEmptyExportFileSystemState' descriptor. 2025-01-08T02:22:54,523 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=179, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,523 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testEmptyExportFileSystemState' from region states. 2025-01-08T02:22:54,523 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302974523"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:54,523 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736302974523"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:54,525 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:22:54,525 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => abc47aa081596c09e48ebf249fd1a2b2, NAME => 'testtb-testEmptyExportFileSystemState,,1736302960671.abc47aa081596c09e48ebf249fd1a2b2.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 8e7bf6158c7b7488f7c41a1178e903cd, NAME => 'testtb-testEmptyExportFileSystemState,1,1736302960671.8e7bf6158c7b7488f7c41a1178e903cd.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:22:54,525 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testEmptyExportFileSystemState' as deleted. 2025-01-08T02:22:54,525 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testEmptyExportFileSystemState","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736302974525"}]},"ts":"9223372036854775807"} 2025-01-08T02:22:54,526 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testEmptyExportFileSystemState state from META 2025-01-08T02:22:54,528 DEBUG [PEWorker-4 {}] procedure.DeleteTableProcedure(133): Finished pid=179, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState 2025-01-08T02:22:54,529 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=179, state=SUCCESS; DeleteTableProcedure table=testtb-testEmptyExportFileSystemState in 32 msec 2025-01-08T02:22:54,606 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=179 2025-01-08T02:22:54,607 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testEmptyExportFileSystemState, procId: 179 completed 2025-01-08T02:22:54,613 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testEmptyExportFileSystemState" 2025-01-08T02:22:54,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:54,616 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testEmptyExportFileSystemState" 2025-01-08T02:22:54,617 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testEmptyExportFileSystemState 2025-01-08T02:22:54,640 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testEmptyExportFileSystemState Thread=809 (was 792) Potentially hanging thread: HFileArchiver-18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:57564 [Waiting for operation #3] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:52226 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-41 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #12 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:45347 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: Thread-5760 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:43023 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-40 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 11269) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:43851 from appattempt_1736302816589_0007_000001 java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-800197986_1 at /127.0.0.1:35372 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:35388 [Waiting for operation #4] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-39 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-800197986_1 at /127.0.0.1:57552 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-42 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:43023 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) - Thread LEAK? -, OpenFileDescriptor=821 (was 799) - OpenFileDescriptor LEAK? -, MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=560 (was 565), ProcessCount=15 (was 14) - ProcessCount LEAK? -, AvailableMemoryMB=2153 (was 2513) 2025-01-08T02:22:54,640 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=809 is superior to 500 2025-01-08T02:22:54,659 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportWithChecksum Thread=809, OpenFileDescriptor=821, MaxFileDescriptor=1048576, SystemLoadAverage=560, ProcessCount=15, AvailableMemoryMB=2152 2025-01-08T02:22:54,659 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=809 is superior to 500 2025-01-08T02:22:54,660 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:22:54,661 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=180, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:22:54,662 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:22:54,662 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportWithChecksum" procId is: 180 2025-01-08T02:22:54,663 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:22:54,663 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T02:22:54,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742277_1453 (size=440) 2025-01-08T02:22:54,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742277_1453 (size=440) 2025-01-08T02:22:54,670 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742277_1453 (size=440) 2025-01-08T02:22:54,672 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 309344e6ac66af78cd30d2baec21e5b1, NAME => 'testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:54,673 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => a88b28b80a439cdaac59080bea96479e, NAME => 'testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportWithChecksum', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:54,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742279_1455 (size=65) 2025-01-08T02:22:54,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742279_1455 (size=65) 2025-01-08T02:22:54,684 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742279_1455 (size=65) 2025-01-08T02:22:54,685 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:54,685 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1681): Closing a88b28b80a439cdaac59080bea96479e, disabling compactions & flushes 2025-01-08T02:22:54,685 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:54,685 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:54,685 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. after waiting 0 ms 2025-01-08T02:22:54,685 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:54,685 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:54,685 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-1 {}] regionserver.HRegion(1635): Region close journal for a88b28b80a439cdaac59080bea96479e: 2025-01-08T02:22:54,691 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742278_1454 (size=65) 2025-01-08T02:22:54,691 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742278_1454 (size=65) 2025-01-08T02:22:54,692 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742278_1454 (size=65) 2025-01-08T02:22:54,692 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:54,692 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1681): Closing 309344e6ac66af78cd30d2baec21e5b1, disabling compactions & flushes 2025-01-08T02:22:54,692 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:54,692 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:54,692 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. after waiting 0 ms 2025-01-08T02:22:54,692 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:54,692 INFO [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:54,692 DEBUG [RegionOpenAndInit-testtb-testExportWithChecksum-pool-0 {}] regionserver.HRegion(1635): Region close journal for 309344e6ac66af78cd30d2baec21e5b1: 2025-01-08T02:22:54,693 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:22:54,693 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736302974693"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302974693"}]},"ts":"1736302974693"} 2025-01-08T02:22:54,694 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.","families":{"info":[{"qualifier":"regioninfo","vlen":64,"tag":[],"timestamp":"1736302974693"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736302974693"}]},"ts":"1736302974693"} 2025-01-08T02:22:54,695 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:22:54,696 INFO [PEWorker-1 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:22:54,696 DEBUG [PEWorker-1 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302974696"}]},"ts":"1736302974696"} 2025-01-08T02:22:54,697 INFO [PEWorker-1 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=ENABLING in hbase:meta 2025-01-08T02:22:54,700 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:22:54,702 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:22:54,702 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:22:54,702 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:22:54,702 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:22:54,702 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:22:54,702 INFO [PEWorker-1 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:22:54,702 DEBUG [PEWorker-1 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:22:54,702 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=309344e6ac66af78cd30d2baec21e5b1, ASSIGN}, {pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=a88b28b80a439cdaac59080bea96479e, ASSIGN}] 2025-01-08T02:22:54,703 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=a88b28b80a439cdaac59080bea96479e, ASSIGN 2025-01-08T02:22:54,703 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=309344e6ac66af78cd30d2baec21e5b1, ASSIGN 2025-01-08T02:22:54,703 INFO [PEWorker-2 {}] assignment.TransitRegionStateProcedure(264): Starting pid=181, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=309344e6ac66af78cd30d2baec21e5b1, ASSIGN; state=OFFLINE, location=c4e5d52b9759,37031,1736302808764; forceNewPlan=false, retain=false 2025-01-08T02:22:54,703 INFO [PEWorker-5 {}] assignment.TransitRegionStateProcedure(264): Starting pid=182, ppid=180, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=a88b28b80a439cdaac59080bea96479e, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:22:54,764 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T02:22:54,854 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:22:54,854 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=182 updating hbase:meta row=a88b28b80a439cdaac59080bea96479e, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:54,854 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=181 updating hbase:meta row=309344e6ac66af78cd30d2baec21e5b1, regionState=OPENING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:54,855 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=183, ppid=182, state=RUNNABLE; OpenRegionProcedure a88b28b80a439cdaac59080bea96479e, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:22:54,856 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=184, ppid=181, state=RUNNABLE; OpenRegionProcedure 309344e6ac66af78cd30d2baec21e5b1, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:22:54,964 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T02:22:55,007 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:55,008 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:55,009 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] handler.AssignRegionHandler(135): Open testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7285): Opening region: {ENCODED => a88b28b80a439cdaac59080bea96479e, NAME => 'testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:22:55,010 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] handler.AssignRegionHandler(135): Open testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7285): Opening region: {ENCODED => 309344e6ac66af78cd30d2baec21e5b1, NAME => 'testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. service=AccessControlService 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. service=AccessControlService 2025-01-08T02:22:55,010 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:55,010 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithChecksum 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportWithChecksum a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:55,010 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(894): Instantiated testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:22:55,011 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7327): checking encryption for 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,011 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7327): checking encryption for a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,011 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(7330): checking classloading for 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,011 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(7330): checking classloading for a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,012 INFO [StoreOpener-309344e6ac66af78cd30d2baec21e5b1-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,012 INFO [StoreOpener-a88b28b80a439cdaac59080bea96479e-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,013 INFO [StoreOpener-309344e6ac66af78cd30d2baec21e5b1-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 309344e6ac66af78cd30d2baec21e5b1 columnFamilyName cf 2025-01-08T02:22:55,013 INFO [StoreOpener-a88b28b80a439cdaac59080bea96479e-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region a88b28b80a439cdaac59080bea96479e columnFamilyName cf 2025-01-08T02:22:55,014 DEBUG [StoreOpener-309344e6ac66af78cd30d2baec21e5b1-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:55,014 DEBUG [StoreOpener-a88b28b80a439cdaac59080bea96479e-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:55,014 INFO [StoreOpener-309344e6ac66af78cd30d2baec21e5b1-1 {}] regionserver.HStore(327): Store=309344e6ac66af78cd30d2baec21e5b1/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:55,014 INFO [StoreOpener-a88b28b80a439cdaac59080bea96479e-1 {}] regionserver.HStore(327): Store=a88b28b80a439cdaac59080bea96479e/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:22:55,015 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,015 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,015 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,015 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,017 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1085): writing seq id for a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,017 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1085): writing seq id for 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,019 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:55,019 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:22:55,019 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1102): Opened a88b28b80a439cdaac59080bea96479e; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=72403651, jitterRate=0.07889847457408905}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:55,019 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1102): Opened 309344e6ac66af78cd30d2baec21e5b1; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=65566336, jitterRate=-0.022985458374023438}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:22:55,019 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegion(1001): Region open journal for 309344e6ac66af78cd30d2baec21e5b1: 2025-01-08T02:22:55,019 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegion(1001): Region open journal for a88b28b80a439cdaac59080bea96479e: 2025-01-08T02:22:55,020 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e., pid=183, masterSystemTime=1736302975007 2025-01-08T02:22:55,020 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1., pid=184, masterSystemTime=1736302975008 2025-01-08T02:22:55,021 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:55,021 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=183}] handler.AssignRegionHandler(164): Opened testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:55,021 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=182 updating hbase:meta row=a88b28b80a439cdaac59080bea96479e, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:55,022 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,022 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=184}] handler.AssignRegionHandler(164): Opened testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,022 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=181 updating hbase:meta row=309344e6ac66af78cd30d2baec21e5b1, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:55,024 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=183, resume processing ppid=182 2025-01-08T02:22:55,024 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=183, ppid=182, state=SUCCESS; OpenRegionProcedure a88b28b80a439cdaac59080bea96479e, server=c4e5d52b9759,39911,1736302808913 in 167 msec 2025-01-08T02:22:55,024 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=184, resume processing ppid=181 2025-01-08T02:22:55,024 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=184, ppid=181, state=SUCCESS; OpenRegionProcedure 309344e6ac66af78cd30d2baec21e5b1, server=c4e5d52b9759,37031,1736302808764 in 167 msec 2025-01-08T02:22:55,025 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=182, ppid=180, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=a88b28b80a439cdaac59080bea96479e, ASSIGN in 322 msec 2025-01-08T02:22:55,026 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=181, resume processing ppid=180 2025-01-08T02:22:55,026 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=181, ppid=180, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=309344e6ac66af78cd30d2baec21e5b1, ASSIGN in 322 msec 2025-01-08T02:22:55,026 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:22:55,026 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736302975026"}]},"ts":"1736302975026"} 2025-01-08T02:22:55,027 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=ENABLED in hbase:meta 2025-01-08T02:22:55,029 INFO [PEWorker-2 {}] procedure.CreateTableProcedure(89): pid=180, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportWithChecksum execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:22:55,029 DEBUG [PEWorker-2 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportWithChecksum jenkins: RWXCA 2025-01-08T02:22:55,031 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T02:22:55,032 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:55,032 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:55,032 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:55,032 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,034 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF\x0AC\x0A\x07jenkins\x128\x08\x03"4\x0A(\x0A\x07default\x12\x1Dtesttb-testExportWithChecksum \x00 \x01 \x02 \x03 \x04 2025-01-08T02:22:55,035 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=180, state=SUCCESS; CreateTableProcedure table=testtb-testExportWithChecksum in 373 msec 2025-01-08T02:22:55,265 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=180 2025-01-08T02:22:55,266 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportWithChecksum, procId: 180 completed 2025-01-08T02:22:55,268 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithChecksum 2025-01-08T02:22:55,268 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,268 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:55,277 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T02:22:55,277 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302975277 (current time:1736302975277). 2025-01-08T02:22:55,277 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:55,277 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportWithChecksum VERSION not specified, setting to 2 2025-01-08T02:22:55,277 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:55,278 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5798cdf7 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@2a98dd01 2025-01-08T02:22:55,281 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7d21136f, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:55,282 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:55,283 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56316, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:55,284 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5798cdf7 to 127.0.0.1:61593 2025-01-08T02:22:55,284 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:55,284 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x32afcdb4 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5eb36e35 2025-01-08T02:22:55,287 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@2bfcb4e7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:55,288 DEBUG [hconnection-0x42860174-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:55,289 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56320, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:55,290 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:55,291 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45744, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:55,292 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x32afcdb4 to 127.0.0.1:61593 2025-01-08T02:22:55,292 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:55,292 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T02:22:55,292 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:55,293 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=185, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T02:22:55,293 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 185 2025-01-08T02:22:55,294 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T02:22:55,294 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:55,295 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:55,297 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:55,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742280_1456 (size=161) 2025-01-08T02:22:55,303 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742280_1456 (size=161) 2025-01-08T02:22:55,304 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742280_1456 (size=161) 2025-01-08T02:22:55,304 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:55,305 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1}, {pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e}] 2025-01-08T02:22:55,305 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,306 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,395 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T02:22:55,456 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:55,456 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:55,457 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=186 2025-01-08T02:22:55,457 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=187 2025-01-08T02:22:55,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:55,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.HRegion(2538): Flush status journal for 309344e6ac66af78cd30d2baec21e5b1: 2025-01-08T02:22:55,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.HRegion(2538): Flush status journal for a88b28b80a439cdaac59080bea96479e: 2025-01-08T02:22:55,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. for emptySnaptb0-testExportWithChecksum completed. 2025-01-08T02:22:55,457 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. for emptySnaptb0-testExportWithChecksum completed. 2025-01-08T02:22:55,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.' region-info for snapshot=emptySnaptb0-testExportWithChecksum 2025-01-08T02:22:55,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:55,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:55,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.' region-info for snapshot=emptySnaptb0-testExportWithChecksum 2025-01-08T02:22:55,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:55,458 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:22:55,464 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742282_1458 (size=68) 2025-01-08T02:22:55,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742281_1457 (size=68) 2025-01-08T02:22:55,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742282_1458 (size=68) 2025-01-08T02:22:55,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742282_1458 (size=68) 2025-01-08T02:22:55,465 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742281_1457 (size=68) 2025-01-08T02:22:55,466 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742281_1457 (size=68) 2025-01-08T02:22:55,466 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:55,466 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=187}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=187 2025-01-08T02:22:55,466 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,466 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=186}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=186 2025-01-08T02:22:55,466 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=186 2025-01-08T02:22:55,466 DEBUG [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=187 2025-01-08T02:22:55,466 INFO [RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithChecksum on region a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,466 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportWithChecksum on region 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,467 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=187, ppid=185, state=RUNNABLE; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,467 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=186, ppid=185, state=RUNNABLE; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,468 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=187, ppid=185, state=SUCCESS; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e in 162 msec 2025-01-08T02:22:55,469 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=186, resume processing ppid=185 2025-01-08T02:22:55,469 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:55,469 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=186, ppid=185, state=SUCCESS; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1 in 162 msec 2025-01-08T02:22:55,470 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:55,471 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:55,471 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:55,471 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:55,471 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:22:55,478 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742283_1459 (size=60) 2025-01-08T02:22:55,478 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742283_1459 (size=60) 2025-01-08T02:22:55,479 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742283_1459 (size=60) 2025-01-08T02:22:55,479 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:55,480 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportWithChecksum 2025-01-08T02:22:55,480 DEBUG [PEWorker-4 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithChecksum 2025-01-08T02:22:55,489 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742284_1460 (size=641) 2025-01-08T02:22:55,490 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742284_1460 (size=641) 2025-01-08T02:22:55,490 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742284_1460 (size=641) 2025-01-08T02:22:55,497 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:55,500 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:55,501 DEBUG [PEWorker-4 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportWithChecksum to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testExportWithChecksum 2025-01-08T02:22:55,502 INFO [PEWorker-4 {}] procedure.SnapshotProcedure(134): pid=185, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:55,502 DEBUG [PEWorker-4 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 185 2025-01-08T02:22:55,503 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=185, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=185, snapshot={ ss=emptySnaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } in 209 msec 2025-01-08T02:22:55,595 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=185 2025-01-08T02:22:55,596 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithChecksum, procId: 185 completed 2025-01-08T02:22:55,598 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=37031 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:55,599 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:22:55,601 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportWithChecksum 2025-01-08T02:22:55,601 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,602 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:22:55,610 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T02:22:55,610 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736302975610 (current time:1736302975610). 2025-01-08T02:22:55,610 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:22:55,610 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportWithChecksum VERSION not specified, setting to 2 2025-01-08T02:22:55,610 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:22:55,611 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x051a7dc5 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@776551b3 2025-01-08T02:22:55,614 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@3c11aefc, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:55,615 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:55,616 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56334, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:55,616 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x051a7dc5 to 127.0.0.1:61593 2025-01-08T02:22:55,617 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:55,617 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x5c14c3bc to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@121f944 2025-01-08T02:22:55,620 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@30f6d1d2, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:22:55,621 DEBUG [hconnection-0x962d995-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:55,622 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:56338, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:55,623 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:22:55,623 INFO [RS-EventLoopGroup-3-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:45746, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:22:55,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5c14c3bc to 127.0.0.1:61593 2025-01-08T02:22:55,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:22:55,624 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportWithChecksum], kv [jenkins: RWXCA] 2025-01-08T02:22:55,625 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:22:55,625 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=188, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } 2025-01-08T02:22:55,626 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 188 2025-01-08T02:22:55,626 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:22:55,626 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T02:22:55,627 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:22:55,629 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:22:55,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742285_1461 (size=156) 2025-01-08T02:22:55,637 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742285_1461 (size=156) 2025-01-08T02:22:55,638 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742285_1461 (size=156) 2025-01-08T02:22:55,639 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:22:55,639 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1}, {pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e}] 2025-01-08T02:22:55,640 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,640 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,727 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T02:22:55,791 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:22:55,791 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:22:55,791 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=37031 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=189 2025-01-08T02:22:55,791 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=190 2025-01-08T02:22:55,791 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,791 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:55,791 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(2837): Flushing 309344e6ac66af78cd30d2baec21e5b1 1/1 column families, dataSize=199 B heapSize=688 B 2025-01-08T02:22:55,791 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(2837): Flushing a88b28b80a439cdaac59080bea96479e 1/1 column families, dataSize=3.06 KB heapSize=6.86 KB 2025-01-08T02:22:55,809 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108882316213c98464bbd4e216579f0fa5f_309344e6ac66af78cd30d2baec21e5b1 is 71, key is 036546a2177713ecc4b4095ffedc9a2f/cf:q/1736302975598/Put/seqid=0 2025-01-08T02:22:55,809 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010893a73f1f282844c588aaca8218502257_a88b28b80a439cdaac59080bea96479e is 71, key is 12c1034d99a729e862a73c749fe14fc9/cf:q/1736302975598/Put/seqid=0 2025-01-08T02:22:55,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742286_1462 (size=5102) 2025-01-08T02:22:55,819 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742286_1462 (size=5102) 2025-01-08T02:22:55,820 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742286_1462 (size=5102) 2025-01-08T02:22:55,820 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:55,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742287_1463 (size=8171) 2025-01-08T02:22:55,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742287_1463 (size=8171) 2025-01-08T02:22:55,821 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742287_1463 (size=8171) 2025-01-08T02:22:55,822 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:55,826 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b2025010893a73f1f282844c588aaca8218502257_a88b28b80a439cdaac59080bea96479e to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b2025010893a73f1f282844c588aaca8218502257_a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,827 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/.tmp/cf/57d3f708e3094de081f1750a305aa4b4, store: [table=testtb-testExportWithChecksum family=cf region=a88b28b80a439cdaac59080bea96479e] 2025-01-08T02:22:55,828 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/.tmp/cf/57d3f708e3094de081f1750a305aa4b4 is 206, key is 1a1485b0369407c1b5a6ffecbeac9c832/cf:q/1736302975598/Put/seqid=0 2025-01-08T02:22:55,831 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108882316213c98464bbd4e216579f0fa5f_309344e6ac66af78cd30d2baec21e5b1 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108882316213c98464bbd4e216579f0fa5f_309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,832 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/.tmp/cf/a42ae286b9d2472ba3ee13971ce93f7d, store: [table=testtb-testExportWithChecksum family=cf region=309344e6ac66af78cd30d2baec21e5b1] 2025-01-08T02:22:55,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742288_1464 (size=14853) 2025-01-08T02:22:55,832 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742288_1464 (size=14853) 2025-01-08T02:22:55,833 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/.tmp/cf/a42ae286b9d2472ba3ee13971ce93f7d is 206, key is 03352c7ed18c1efa64ff1a45bfe47bfe3/cf:q/1736302975598/Put/seqid=0 2025-01-08T02:22:55,833 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742288_1464 (size=14853) 2025-01-08T02:22:55,833 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.1 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/.tmp/cf/57d3f708e3094de081f1750a305aa4b4 2025-01-08T02:22:55,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742289_1465 (size=5906) 2025-01-08T02:22:55,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742289_1465 (size=5906) 2025-01-08T02:22:55,838 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742289_1465 (size=5906) 2025-01-08T02:22:55,838 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=199, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/.tmp/cf/a42ae286b9d2472ba3ee13971ce93f7d 2025-01-08T02:22:55,843 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/.tmp/cf/a42ae286b9d2472ba3ee13971ce93f7d as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/cf/a42ae286b9d2472ba3ee13971ce93f7d 2025-01-08T02:22:55,844 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/.tmp/cf/57d3f708e3094de081f1750a305aa4b4 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4 2025-01-08T02:22:55,847 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/cf/a42ae286b9d2472ba3ee13971ce93f7d, entries=3, sequenceid=6, filesize=5.8 K 2025-01-08T02:22:55,848 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4, entries=47, sequenceid=6, filesize=14.5 K 2025-01-08T02:22:55,848 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(3040): Finished flush of dataSize ~199 B/199, heapSize ~672 B/672, currentSize=0 B/0 for 309344e6ac66af78cd30d2baec21e5b1 in 57ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:55,848 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithChecksum' 2025-01-08T02:22:55,849 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(3040): Finished flush of dataSize ~3.06 KB/3137, heapSize ~6.84 KB/7008, currentSize=0 B/0 for a88b28b80a439cdaac59080bea96479e in 58ms, sequenceid=6, compaction requested=false 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportWithChecksum' 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.HRegion(2538): Flush status journal for a88b28b80a439cdaac59080bea96479e: 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.HRegion(2538): Flush status journal for 309344e6ac66af78cd30d2baec21e5b1: 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. for snaptb0-testExportWithChecksum completed. 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. for snaptb0-testExportWithChecksum completed. 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.' region-info for snapshot=snaptb0-testExportWithChecksum 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4] hfiles 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.' region-info for snapshot=snaptb0-testExportWithChecksum 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4 for snapshot=snaptb0-testExportWithChecksum 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/cf/a42ae286b9d2472ba3ee13971ce93f7d] hfiles 2025-01-08T02:22:55,849 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/cf/a42ae286b9d2472ba3ee13971ce93f7d for snapshot=snaptb0-testExportWithChecksum 2025-01-08T02:22:55,855 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742290_1466 (size=107) 2025-01-08T02:22:55,855 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742290_1466 (size=107) 2025-01-08T02:22:55,855 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742290_1466 (size=107) 2025-01-08T02:22:55,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:22:55,855 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=190}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=190 2025-01-08T02:22:55,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=190 2025-01-08T02:22:55,856 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithChecksum on region a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,856 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=190, ppid=188, state=RUNNABLE; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742291_1467 (size=107) 2025-01-08T02:22:55,858 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=190, ppid=188, state=SUCCESS; SnapshotRegionProcedure a88b28b80a439cdaac59080bea96479e in 218 msec 2025-01-08T02:22:55,858 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742291_1467 (size=107) 2025-01-08T02:22:55,859 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742291_1467 (size=107) 2025-01-08T02:22:55,859 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:22:55,859 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=189}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=189 2025-01-08T02:22:55,859 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=189 2025-01-08T02:22:55,859 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportWithChecksum on region 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,859 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=189, ppid=188, state=RUNNABLE; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,861 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=189, resume processing ppid=188 2025-01-08T02:22:55,861 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=189, ppid=188, state=SUCCESS; SnapshotRegionProcedure 309344e6ac66af78cd30d2baec21e5b1 in 221 msec 2025-01-08T02:22:55,861 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:22:55,862 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:22:55,862 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:22:55,862 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:22:55,862 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:22:55,863 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b2025010893a73f1f282844c588aaca8218502257_a88b28b80a439cdaac59080bea96479e, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108882316213c98464bbd4e216579f0fa5f_309344e6ac66af78cd30d2baec21e5b1] hfiles 2025-01-08T02:22:55,863 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b2025010893a73f1f282844c588aaca8218502257_a88b28b80a439cdaac59080bea96479e 2025-01-08T02:22:55,863 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108882316213c98464bbd4e216579f0fa5f_309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:22:55,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742292_1468 (size=291) 2025-01-08T02:22:55,868 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742292_1468 (size=291) 2025-01-08T02:22:55,869 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742292_1468 (size=291) 2025-01-08T02:22:55,869 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:22:55,869 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportWithChecksum 2025-01-08T02:22:55,870 DEBUG [PEWorker-5 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T02:22:55,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742293_1469 (size=951) 2025-01-08T02:22:55,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742293_1469 (size=951) 2025-01-08T02:22:55,881 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742293_1469 (size=951) 2025-01-08T02:22:55,885 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:22:55,890 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:22:55,890 DEBUG [PEWorker-5 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T02:22:55,891 INFO [PEWorker-5 {}] procedure.SnapshotProcedure(134): pid=188, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:22:55,891 DEBUG [PEWorker-5 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 }, snapshot procedure id = 188 2025-01-08T02:22:55,892 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=188, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=188, snapshot={ ss=snaptb0-testExportWithChecksum table=testtb-testExportWithChecksum type=FLUSH ttl=0 } in 266 msec 2025-01-08T02:22:55,927 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=188 2025-01-08T02:22:55,928 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportWithChecksum, procId: 188 completed 2025-01-08T02:22:55,928 INFO [Time-limited test {}] snapshot.TestExportSnapshot(476): Local export destination path: /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928 2025-01-08T02:22:55,928 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=file:///, tgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928, rawTgtDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:55,956 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:22:55,956 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=org.apache.hadoop.fs.LocalFileSystem@65f64b18, outputRoot=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928, skipTmp=false, initialOutputSnapshotDir=file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T02:22:55,958 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:22:55,961 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithChecksum to file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T02:22:55,991 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:55,992 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:55,992 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:55,992 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:56,996 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-5116722599558488469.jar 2025-01-08T02:22:56,997 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:56,997 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:57,069 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-6009558726115738591.jar 2025-01-08T02:22:57,069 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:57,069 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:57,070 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:57,070 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:57,070 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:57,070 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:22:57,070 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:22:57,070 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:22:57,071 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:22:57,071 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:22:57,071 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:22:57,071 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:22:57,072 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:22:57,072 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:22:57,072 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:22:57,072 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:22:57,072 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:22:57,073 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:22:57,073 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:57,074 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:57,074 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:57,074 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:57,075 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:22:57,075 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:57,075 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:22:57,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742294_1470 (size=127628) 2025-01-08T02:22:57,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742294_1470 (size=127628) 2025-01-08T02:22:57,132 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742294_1470 (size=127628) 2025-01-08T02:22:57,151 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T02:22:57,152 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T02:22:57,152 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742295_1471 (size=2172137) 2025-01-08T02:22:57,159 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742296_1472 (size=213228) 2025-01-08T02:22:57,159 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742296_1472 (size=213228) 2025-01-08T02:22:57,160 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742296_1472 (size=213228) 2025-01-08T02:22:57,170 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T02:22:57,171 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T02:22:57,171 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742297_1473 (size=1877034) 2025-01-08T02:22:57,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742298_1474 (size=533455) 2025-01-08T02:22:57,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742298_1474 (size=533455) 2025-01-08T02:22:57,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742298_1474 (size=533455) 2025-01-08T02:22:57,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T02:22:57,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T02:22:57,206 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742299_1475 (size=7280644) 2025-01-08T02:22:57,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742300_1476 (size=6350715) 2025-01-08T02:22:57,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742300_1476 (size=6350715) 2025-01-08T02:22:57,229 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742300_1476 (size=6350715) 2025-01-08T02:22:57,247 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742301_1477 (size=4188619) 2025-01-08T02:22:57,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742301_1477 (size=4188619) 2025-01-08T02:22:57,248 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742301_1477 (size=4188619) 2025-01-08T02:22:57,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742302_1478 (size=20406) 2025-01-08T02:22:57,254 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742302_1478 (size=20406) 2025-01-08T02:22:57,255 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742302_1478 (size=20406) 2025-01-08T02:22:57,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742303_1479 (size=75495) 2025-01-08T02:22:57,261 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742303_1479 (size=75495) 2025-01-08T02:22:57,261 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742303_1479 (size=75495) 2025-01-08T02:22:57,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742304_1480 (size=45609) 2025-01-08T02:22:57,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742304_1480 (size=45609) 2025-01-08T02:22:57,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742304_1480 (size=45609) 2025-01-08T02:22:57,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742305_1481 (size=110084) 2025-01-08T02:22:57,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742305_1481 (size=110084) 2025-01-08T02:22:57,273 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742305_1481 (size=110084) 2025-01-08T02:22:57,282 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T02:22:57,282 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T02:22:57,283 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742306_1482 (size=1323991) 2025-01-08T02:22:57,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742307_1483 (size=23076) 2025-01-08T02:22:57,288 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742307_1483 (size=23076) 2025-01-08T02:22:57,289 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742307_1483 (size=23076) 2025-01-08T02:22:57,294 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742308_1484 (size=126803) 2025-01-08T02:22:57,294 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742308_1484 (size=126803) 2025-01-08T02:22:57,295 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742308_1484 (size=126803) 2025-01-08T02:22:57,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742309_1485 (size=322274) 2025-01-08T02:22:57,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742309_1485 (size=322274) 2025-01-08T02:22:57,301 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742309_1485 (size=322274) 2025-01-08T02:22:57,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742310_1486 (size=1832290) 2025-01-08T02:22:57,311 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742310_1486 (size=1832290) 2025-01-08T02:22:57,312 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742310_1486 (size=1832290) 2025-01-08T02:22:57,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742311_1487 (size=30081) 2025-01-08T02:22:57,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742311_1487 (size=30081) 2025-01-08T02:22:57,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742311_1487 (size=30081) 2025-01-08T02:22:57,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742312_1488 (size=53616) 2025-01-08T02:22:57,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742312_1488 (size=53616) 2025-01-08T02:22:57,324 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742312_1488 (size=53616) 2025-01-08T02:22:57,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742313_1489 (size=29229) 2025-01-08T02:22:57,330 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742313_1489 (size=29229) 2025-01-08T02:22:57,331 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742313_1489 (size=29229) 2025-01-08T02:22:57,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742314_1490 (size=169089) 2025-01-08T02:22:57,338 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742314_1490 (size=169089) 2025-01-08T02:22:57,339 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742314_1490 (size=169089) 2025-01-08T02:22:57,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742315_1491 (size=5175431) 2025-01-08T02:22:57,362 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742315_1491 (size=5175431) 2025-01-08T02:22:57,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742315_1491 (size=5175431) 2025-01-08T02:22:57,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742316_1492 (size=136454) 2025-01-08T02:22:57,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742316_1492 (size=136454) 2025-01-08T02:22:57,371 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742316_1492 (size=136454) 2025-01-08T02:22:57,379 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742317_1493 (size=451756) 2025-01-08T02:22:57,380 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742317_1493 (size=451756) 2025-01-08T02:22:57,380 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742317_1493 (size=451756) 2025-01-08T02:22:57,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742318_1494 (size=907469) 2025-01-08T02:22:57,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742318_1494 (size=907469) 2025-01-08T02:22:57,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742318_1494 (size=907469) 2025-01-08T02:22:57,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T02:22:57,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T02:22:57,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742319_1495 (size=3317408) 2025-01-08T02:22:57,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742320_1496 (size=503880) 2025-01-08T02:22:57,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742320_1496 (size=503880) 2025-01-08T02:22:57,417 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742320_1496 (size=503880) 2025-01-08T02:22:57,439 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T02:22:57,439 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T02:22:57,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742321_1497 (size=4695811) 2025-01-08T02:22:57,441 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:22:57,443 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithChecksum' hfile list 2025-01-08T02:22:57,445 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T02:22:57,451 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742322_1498 (size=714) 2025-01-08T02:22:57,451 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742322_1498 (size=714) 2025-01-08T02:22:57,452 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742322_1498 (size=714) 2025-01-08T02:22:57,459 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742323_1499 (size=15) 2025-01-08T02:22:57,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742323_1499 (size=15) 2025-01-08T02:22:57,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742323_1499 (size=15) 2025-01-08T02:22:57,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742324_1500 (size=304925) 2025-01-08T02:22:57,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742324_1500 (size=304925) 2025-01-08T02:22:57,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742324_1500 (size=304925) 2025-01-08T02:22:58,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum 2025-01-08T02:22:58,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum Metrics about Tables on a single HBase RegionServer 2025-01-08T02:22:58,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testEmptyExportFileSystemState 2025-01-08T02:22:58,714 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:22:58,714 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:22:58,717 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0007_000001 (auth:SIMPLE) from 127.0.0.1:36814 2025-01-08T02:22:58,728 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0007/container_1736302816589_0007_01_000001/launch_container.sh] 2025-01-08T02:22:58,728 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0007/container_1736302816589_0007_01_000001/container_tokens] 2025-01-08T02:22:58,728 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0007/container_1736302816589_0007_01_000001/sysfs] 2025-01-08T02:22:59,643 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:52268 2025-01-08T02:22:59,864 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:23:04,625 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:53412 2025-01-08T02:23:04,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742325_1501 (size=350599) 2025-01-08T02:23:04,851 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742325_1501 (size=350599) 2025-01-08T02:23:04,852 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742325_1501 (size=350599) 2025-01-08T02:23:06,390 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:23:06,852 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:42756 2025-01-08T02:23:10,568 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000002/launch_container.sh] 2025-01-08T02:23:10,568 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000002/container_tokens] 2025-01-08T02:23:10,568 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000002/sysfs] Error: java.io.IOException: Checksum mismatch between hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4 and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928/archive/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T02:23:11,716 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:42770 2025-01-08T02:23:12,132 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 2f4301e371bc30413722a847ac142fb1 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:23:12,133 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 309344e6ac66af78cd30d2baec21e5b1 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:23:12,133 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region 4c7db76a9413d28afb5a4d977c0db552 changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:23:12,133 DEBUG [master/c4e5d52b9759:0.Chore.1 {}] balancer.RegionLocationFinder(172): Locality for region a88b28b80a439cdaac59080bea96479e changed from -1.0 to 0.0, refreshing cache 2025-01-08T02:23:14,157 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 2f4301e371bc30413722a847ac142fb1, had cached 0 bytes from a total of 6086 2025-01-08T02:23:14,158 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 4c7db76a9413d28afb5a4d977c0db552, had cached 0 bytes from a total of 14463 2025-01-08T02:23:15,203 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000003/launch_container.sh] 2025-01-08T02:23:15,203 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000003/container_tokens] 2025-01-08T02:23:15,203 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000003/sysfs] Error: java.io.IOException: Checksum mismatch between hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4 and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928/archive/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T02:23:16,729 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:46906 2025-01-08T02:23:19,984 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000004/launch_container.sh] 2025-01-08T02:23:19,984 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000004/container_tokens] 2025-01-08T02:23:19,984 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000004/sysfs] Error: java.io.IOException: Checksum mismatch between hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4 and file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/local-export-1736302975928/archive/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4. Input and output filesystems are of different types. Their checksum algorithms may be incompatible. You can choose file-level checksum validation via -Ddfs.checksum.combine.mode=COMPOSITE_CRC when block-sizes or filesystems are different. Or you can skip checksum-checks altogether with -no-checksum-verify, for the table backup scenario, you should use -i option to skip checksum-checks. (NOTE: By skipping checksums, one runs the risk of masking data-corruption during file-transfer.) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.verifyCopyResult(ExportSnapshot.java:596) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.copyFile(ExportSnapshot.java:332) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:254) at org.apache.hadoop.hbase.snapshot.ExportSnapshot$ExportMapper.map(ExportSnapshot.java:180) at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145) at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:800) at org.apache.hadoop.mapred.MapTask.run(MapTask.java:348) at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:178) at java.base/java.security.AccessController.doPrivileged(AccessController.java:712) at java.base/javax.security.auth.Subject.doAs(Subject.java:439) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:172) 2025-01-08T02:23:21,750 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:46914 2025-01-08T02:23:25,750 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742326_1502 (size=21340) 2025-01-08T02:23:25,750 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742326_1502 (size=21340) 2025-01-08T02:23:25,750 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742326_1502 (size=21340) 2025-01-08T02:23:25,823 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742327_1503 (size=460) 2025-01-08T02:23:25,823 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742327_1503 (size=460) 2025-01-08T02:23:25,824 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742327_1503 (size=460) 2025-01-08T02:23:25,868 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000005/launch_container.sh] 2025-01-08T02:23:25,868 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000005/container_tokens] 2025-01-08T02:23:25,869 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_0/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000005/sysfs] 2025-01-08T02:23:25,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742328_1504 (size=21340) 2025-01-08T02:23:25,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742328_1504 (size=21340) 2025-01-08T02:23:25,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742328_1504 (size=21340) 2025-01-08T02:23:25,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742329_1505 (size=350599) 2025-01-08T02:23:25,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742329_1505 (size=350599) 2025-01-08T02:23:25,940 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742329_1505 (size=350599) 2025-01-08T02:23:25,955 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:55006 2025-01-08T02:23:27,926 ERROR [Time-limited test {}] snapshot.ExportSnapshot(1227): Snapshot export failed org.apache.hadoop.hbase.snapshot.ExportSnapshotException: Task failed task_1736302816589_0008_m_000000 Job failed as tasks failed. failedMaps:1 failedReduces:0 killedMaps:0 killedReduces: 0 at org.apache.hadoop.hbase.snapshot.ExportSnapshot.runCopyJob(ExportSnapshot.java:935) ~[classes/:?] at org.apache.hadoop.hbase.snapshot.ExportSnapshot.doWork(ExportSnapshot.java:1204) ~[classes/:?] at org.apache.hadoop.hbase.util.AbstractHBaseTool.run(AbstractHBaseTool.java:151) ~[hbase-common-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:82) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.runExportSnapshot(TestExportSnapshot.java:523) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportFileSystemState(TestExportSnapshot.java:353) ~[test-classes/:?] at org.apache.hadoop.hbase.snapshot.TestExportSnapshot.testExportWithChecksum(TestExportSnapshot.java:237) ~[test-classes/:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) ~[junit-4.13.2.jar:4.13.2] at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299) ~[junit-4.13.2.jar:4.13.2] at org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293) ~[junit-4.13.2.jar:4.13.2] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:23:27,928 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927 2025-01-08T02:23:27,928 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:23:27,957 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:23:27,957 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927, skipTmp=false, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T02:23:27,959 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:23:27,964 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithChecksum to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927/.hbase-snapshot/.tmp/snaptb0-testExportWithChecksum 2025-01-08T02:23:27,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742330_1506 (size=156) 2025-01-08T02:23:27,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742331_1507 (size=951) 2025-01-08T02:23:27,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742330_1506 (size=156) 2025-01-08T02:23:27,978 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742331_1507 (size=951) 2025-01-08T02:23:27,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742331_1507 (size=951) 2025-01-08T02:23:27,979 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742330_1506 (size=156) 2025-01-08T02:23:27,980 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:27,981 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:27,981 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:27,981 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,110 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-5837779996275958285.jar 2025-01-08T02:23:29,110 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,110 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,180 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-7980155334706211912.jar 2025-01-08T02:23:29,180 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,180 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,180 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:29,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:23:29,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:23:29,181 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:23:29,182 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:23:29,182 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:23:29,182 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:23:29,182 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:23:29,182 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:23:29,183 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:23:29,183 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:23:29,183 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:23:29,183 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:23:29,183 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:29,184 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:29,184 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:23:29,184 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:29,185 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:29,185 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:23:29,185 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:23:29,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742332_1508 (size=127628) 2025-01-08T02:23:29,238 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742332_1508 (size=127628) 2025-01-08T02:23:29,239 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742332_1508 (size=127628) 2025-01-08T02:23:29,251 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T02:23:29,251 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T02:23:29,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742333_1509 (size=2172137) 2025-01-08T02:23:29,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742334_1510 (size=213228) 2025-01-08T02:23:29,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742334_1510 (size=213228) 2025-01-08T02:23:29,258 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742334_1510 (size=213228) 2025-01-08T02:23:29,274 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T02:23:29,275 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T02:23:29,275 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742335_1511 (size=1877034) 2025-01-08T02:23:29,283 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742336_1512 (size=533455) 2025-01-08T02:23:29,283 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742336_1512 (size=533455) 2025-01-08T02:23:29,284 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742336_1512 (size=533455) 2025-01-08T02:23:29,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T02:23:29,317 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T02:23:29,318 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742337_1513 (size=7280644) 2025-01-08T02:23:29,335 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T02:23:29,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T02:23:29,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742338_1514 (size=4188619) 2025-01-08T02:23:29,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742339_1515 (size=20406) 2025-01-08T02:23:29,344 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742339_1515 (size=20406) 2025-01-08T02:23:29,345 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742339_1515 (size=20406) 2025-01-08T02:23:29,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742340_1516 (size=75495) 2025-01-08T02:23:29,351 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742340_1516 (size=75495) 2025-01-08T02:23:29,352 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742340_1516 (size=75495) 2025-01-08T02:23:29,358 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742341_1517 (size=45609) 2025-01-08T02:23:29,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742341_1517 (size=45609) 2025-01-08T02:23:29,359 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742341_1517 (size=45609) 2025-01-08T02:23:29,366 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742342_1518 (size=110084) 2025-01-08T02:23:29,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742342_1518 (size=110084) 2025-01-08T02:23:29,367 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742342_1518 (size=110084) 2025-01-08T02:23:29,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742343_1519 (size=451756) 2025-01-08T02:23:29,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742343_1519 (size=451756) 2025-01-08T02:23:29,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742343_1519 (size=451756) 2025-01-08T02:23:29,385 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742344_1520 (size=1323991) 2025-01-08T02:23:29,386 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742344_1520 (size=1323991) 2025-01-08T02:23:29,386 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742344_1520 (size=1323991) 2025-01-08T02:23:29,393 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742345_1521 (size=23076) 2025-01-08T02:23:29,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742345_1521 (size=23076) 2025-01-08T02:23:29,394 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742345_1521 (size=23076) 2025-01-08T02:23:29,400 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742346_1522 (size=126803) 2025-01-08T02:23:29,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742346_1522 (size=126803) 2025-01-08T02:23:29,401 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742346_1522 (size=126803) 2025-01-08T02:23:29,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742347_1523 (size=322274) 2025-01-08T02:23:29,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742347_1523 (size=322274) 2025-01-08T02:23:29,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742347_1523 (size=322274) 2025-01-08T02:23:29,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742348_1524 (size=1832290) 2025-01-08T02:23:29,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742348_1524 (size=1832290) 2025-01-08T02:23:29,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742348_1524 (size=1832290) 2025-01-08T02:23:29,427 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742349_1525 (size=30081) 2025-01-08T02:23:29,427 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742349_1525 (size=30081) 2025-01-08T02:23:29,428 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742349_1525 (size=30081) 2025-01-08T02:23:29,433 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742350_1526 (size=53616) 2025-01-08T02:23:29,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742350_1526 (size=53616) 2025-01-08T02:23:29,434 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742350_1526 (size=53616) 2025-01-08T02:23:29,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742351_1527 (size=29229) 2025-01-08T02:23:29,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742351_1527 (size=29229) 2025-01-08T02:23:29,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742351_1527 (size=29229) 2025-01-08T02:23:29,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742352_1528 (size=169089) 2025-01-08T02:23:29,447 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742352_1528 (size=169089) 2025-01-08T02:23:29,448 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742352_1528 (size=169089) 2025-01-08T02:23:29,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742353_1529 (size=6350715) 2025-01-08T02:23:29,476 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742353_1529 (size=6350715) 2025-01-08T02:23:29,477 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742353_1529 (size=6350715) 2025-01-08T02:23:29,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T02:23:29,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T02:23:29,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742354_1530 (size=5175431) 2025-01-08T02:23:29,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742355_1531 (size=136454) 2025-01-08T02:23:29,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742355_1531 (size=136454) 2025-01-08T02:23:29,529 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742355_1531 (size=136454) 2025-01-08T02:23:29,540 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742356_1532 (size=907469) 2025-01-08T02:23:29,540 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742356_1532 (size=907469) 2025-01-08T02:23:29,540 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742356_1532 (size=907469) 2025-01-08T02:23:29,557 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T02:23:29,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T02:23:29,558 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742357_1533 (size=3317408) 2025-01-08T02:23:29,565 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742358_1534 (size=503880) 2025-01-08T02:23:29,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742358_1534 (size=503880) 2025-01-08T02:23:29,566 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742358_1534 (size=503880) 2025-01-08T02:23:29,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T02:23:29,586 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T02:23:29,587 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742359_1535 (size=4695811) 2025-01-08T02:23:29,588 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:23:29,590 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportWithChecksum' hfile list 2025-01-08T02:23:29,593 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.2 K 2025-01-08T02:23:29,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742360_1536 (size=714) 2025-01-08T02:23:29,606 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742360_1536 (size=714) 2025-01-08T02:23:29,607 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742360_1536 (size=714) 2025-01-08T02:23:29,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742361_1537 (size=15) 2025-01-08T02:23:29,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742361_1537 (size=15) 2025-01-08T02:23:29,622 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742361_1537 (size=15) 2025-01-08T02:23:29,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742362_1538 (size=304875) 2025-01-08T02:23:29,639 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742362_1538 (size=304875) 2025-01-08T02:23:29,640 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742362_1538 (size=304875) 2025-01-08T02:23:30,065 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x58e22cc5b42a90 with lease ID 0x1d06c9fc3b280c51: from storage DS-9dba95bc-92bf-4076-a3ca-0eebd7332f9e node DatanodeRegistration(127.0.0.1:41199, datanodeUuid=5f73ff73-0f7b-4bdf-9032-a33a82e289db, infoPort=41865, infoSecurePort=0, ipcPort=36113, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 71, hasStaleStorage: false, processing time: 1 msecs, invalidatedBlocks: 17 2025-01-08T02:23:30,065 INFO [Block report processor {}] blockmanagement.BlockManager(2972): BLOCK* processReport 0x58e22cc5b42a90 with lease ID 0x1d06c9fc3b280c51: from storage DS-c01f0001-ea9e-4e51-b8a4-76247c985dca node DatanodeRegistration(127.0.0.1:41199, datanodeUuid=5f73ff73-0f7b-4bdf-9032-a33a82e289db, infoPort=41865, infoSecurePort=0, ipcPort=36113, storageInfo=lv=-57;cid=testClusterID;nsid=1261789929;c=1736302801782), blocks: 73, hasStaleStorage: false, processing time: 0 msecs, invalidatedBlocks: 16 2025-01-08T02:23:32,030 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:23:32,030 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:23:32,034 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0008_000001 (auth:SIMPLE) from 127.0.0.1:48078 2025-01-08T02:23:32,050 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000001/launch_container.sh] 2025-01-08T02:23:32,050 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000001/container_tokens] 2025-01-08T02:23:32,050 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_2/usercache/jenkins/appcache/application_1736302816589_0008/container_1736302816589_0008_01_000001/sysfs] 2025-01-08T02:23:32,827 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0009_000001 (auth:SIMPLE) from 127.0.0.1:55020 2025-01-08T02:23:36,390 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:23:37,703 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0009_000001 (auth:SIMPLE) from 127.0.0.1:45656 2025-01-08T02:23:37,950 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742363_1539 (size=350549) 2025-01-08T02:23:37,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742363_1539 (size=350549) 2025-01-08T02:23:37,951 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742363_1539 (size=350549) 2025-01-08T02:23:39,932 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0009_000001 (auth:SIMPLE) from 127.0.0.1:37246 2025-01-08T02:23:40,011 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region a88b28b80a439cdaac59080bea96479e, had cached 0 bytes from a total of 14853 2025-01-08T02:23:40,011 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 309344e6ac66af78cd30d2baec21e5b1, had cached 0 bytes from a total of 5906 2025-01-08T02:23:42,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742364_1540 (size=14853) 2025-01-08T02:23:42,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742364_1540 (size=14853) 2025-01-08T02:23:42,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742364_1540 (size=14853) 2025-01-08T02:23:42,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742365_1541 (size=8171) 2025-01-08T02:23:42,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742365_1541 (size=8171) 2025-01-08T02:23:42,960 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742365_1541 (size=8171) 2025-01-08T02:23:43,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742366_1542 (size=5906) 2025-01-08T02:23:43,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742366_1542 (size=5906) 2025-01-08T02:23:43,004 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742366_1542 (size=5906) 2025-01-08T02:23:43,047 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742367_1543 (size=5102) 2025-01-08T02:23:43,048 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742367_1543 (size=5102) 2025-01-08T02:23:43,048 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742367_1543 (size=5102) 2025-01-08T02:23:43,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742368_1544 (size=17459) 2025-01-08T02:23:43,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742368_1544 (size=17459) 2025-01-08T02:23:43,135 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742368_1544 (size=17459) 2025-01-08T02:23:43,149 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742369_1545 (size=462) 2025-01-08T02:23:43,149 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742369_1545 (size=462) 2025-01-08T02:23:43,150 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742369_1545 (size=462) 2025-01-08T02:23:43,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742370_1546 (size=17459) 2025-01-08T02:23:43,179 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742370_1546 (size=17459) 2025-01-08T02:23:43,180 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742370_1546 (size=17459) 2025-01-08T02:23:43,198 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742371_1547 (size=350549) 2025-01-08T02:23:43,198 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742371_1547 (size=350549) 2025-01-08T02:23:43,199 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742371_1547 (size=350549) 2025-01-08T02:23:43,216 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0009_000001 (auth:SIMPLE) from 127.0.0.1:37262 2025-01-08T02:23:43,239 WARN [ContainersLauncher #3 {}] nodemanager.DefaultContainerExecutor(360): Exit code from container container_1736302816589_0009_01_000002 is : 143 2025-01-08T02:23:43,254 WARN [ContainersLauncher #4 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0009/container_1736302816589_0009_01_000002/launch_container.sh] 2025-01-08T02:23:43,254 WARN [ContainersLauncher #4 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0009/container_1736302816589_0009_01_000002/container_tokens] 2025-01-08T02:23:43,254 WARN [ContainersLauncher #4 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_2/usercache/jenkins/appcache/application_1736302816589_0009/container_1736302816589_0009_01_000002/sysfs] 2025-01-08T02:23:44,775 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:23:44,776 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:23:44,781 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportWithChecksum 2025-01-08T02:23:44,781 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:23:44,782 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:23:44,782 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithChecksum at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T02:23:44,782 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithChecksum/.snapshotinfo 2025-01-08T02:23:44,782 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportWithChecksum/data.manifest 2025-01-08T02:23:44,782 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927/.hbase-snapshot/snaptb0-testExportWithChecksum at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927/.hbase-snapshot/snaptb0-testExportWithChecksum 2025-01-08T02:23:44,783 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927/.hbase-snapshot/snaptb0-testExportWithChecksum/.snapshotinfo 2025-01-08T02:23:44,783 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303007927/.hbase-snapshot/snaptb0-testExportWithChecksum/data.manifest 2025-01-08T02:23:44,787 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportWithChecksum 2025-01-08T02:23:44,787 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportWithChecksum 2025-01-08T02:23:44,788 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=191, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:23:44,789 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T02:23:44,789 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736303024789"}]},"ts":"1736303024789"} 2025-01-08T02:23:44,790 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=DISABLING in hbase:meta 2025-01-08T02:23:44,792 INFO [PEWorker-3 {}] procedure.DisableTableProcedure(284): Set testtb-testExportWithChecksum to state=DISABLING 2025-01-08T02:23:44,793 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=192, ppid=191, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportWithChecksum}] 2025-01-08T02:23:44,794 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=193, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=309344e6ac66af78cd30d2baec21e5b1, UNASSIGN}, {pid=194, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=a88b28b80a439cdaac59080bea96479e, UNASSIGN}] 2025-01-08T02:23:44,794 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=194, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=a88b28b80a439cdaac59080bea96479e, UNASSIGN 2025-01-08T02:23:44,795 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=193, ppid=192, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=309344e6ac66af78cd30d2baec21e5b1, UNASSIGN 2025-01-08T02:23:44,795 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=194 updating hbase:meta row=a88b28b80a439cdaac59080bea96479e, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:23:44,795 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=193 updating hbase:meta row=309344e6ac66af78cd30d2baec21e5b1, regionState=CLOSING, regionLocation=c4e5d52b9759,37031,1736302808764 2025-01-08T02:23:44,796 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:23:44,796 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=195, ppid=194, state=RUNNABLE; CloseRegionProcedure a88b28b80a439cdaac59080bea96479e, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:23:44,797 DEBUG [PEWorker-4 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:23:44,797 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=196, ppid=193, state=RUNNABLE; CloseRegionProcedure 309344e6ac66af78cd30d2baec21e5b1, server=c4e5d52b9759,37031,1736302808764}] 2025-01-08T02:23:44,890 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T02:23:44,948 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:23:44,948 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(124): Close a88b28b80a439cdaac59080bea96479e 2025-01-08T02:23:44,948 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:23:44,948 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1681): Closing a88b28b80a439cdaac59080bea96479e, disabling compactions & flushes 2025-01-08T02:23:44,948 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:23:44,948 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:23:44,948 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. after waiting 0 ms 2025-01-08T02:23:44,948 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:23:44,949 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,37031,1736302808764 2025-01-08T02:23:44,949 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(124): Close 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:23:44,949 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:23:44,949 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1681): Closing 309344e6ac66af78cd30d2baec21e5b1, disabling compactions & flushes 2025-01-08T02:23:44,949 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1703): Closing region testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:23:44,949 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:23:44,949 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. after waiting 0 ms 2025-01-08T02:23:44,949 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:23:44,952 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:23:44,952 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:23:44,952 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:23:44,953 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e. 2025-01-08T02:23:44,953 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] regionserver.HRegion(1635): Region close journal for a88b28b80a439cdaac59080bea96479e: 2025-01-08T02:23:44,953 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:23:44,953 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1922): Closed testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1. 2025-01-08T02:23:44,953 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] regionserver.HRegion(1635): Region close journal for 309344e6ac66af78cd30d2baec21e5b1: 2025-01-08T02:23:44,954 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=195}] handler.UnassignRegionHandler(170): Closed a88b28b80a439cdaac59080bea96479e 2025-01-08T02:23:44,954 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=194 updating hbase:meta row=a88b28b80a439cdaac59080bea96479e, regionState=CLOSED 2025-01-08T02:23:44,954 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=196}] handler.UnassignRegionHandler(170): Closed 309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:23:44,954 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=193 updating hbase:meta row=309344e6ac66af78cd30d2baec21e5b1, regionState=CLOSED 2025-01-08T02:23:44,956 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=195, resume processing ppid=194 2025-01-08T02:23:44,957 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=196, resume processing ppid=193 2025-01-08T02:23:44,957 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=196, ppid=193, state=SUCCESS; CloseRegionProcedure 309344e6ac66af78cd30d2baec21e5b1, server=c4e5d52b9759,37031,1736302808764 in 159 msec 2025-01-08T02:23:44,957 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=195, ppid=194, state=SUCCESS; CloseRegionProcedure a88b28b80a439cdaac59080bea96479e, server=c4e5d52b9759,39911,1736302808913 in 159 msec 2025-01-08T02:23:44,957 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=194, ppid=192, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=a88b28b80a439cdaac59080bea96479e, UNASSIGN in 162 msec 2025-01-08T02:23:44,958 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=193, resume processing ppid=192 2025-01-08T02:23:44,958 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=193, ppid=192, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportWithChecksum, region=309344e6ac66af78cd30d2baec21e5b1, UNASSIGN in 163 msec 2025-01-08T02:23:44,960 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=192, resume processing ppid=191 2025-01-08T02:23:44,960 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=192, ppid=191, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportWithChecksum in 166 msec 2025-01-08T02:23:44,960 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736303024960"}]},"ts":"1736303024960"} 2025-01-08T02:23:44,961 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportWithChecksum, state=DISABLED in hbase:meta 2025-01-08T02:23:44,963 INFO [PEWorker-2 {}] procedure.DisableTableProcedure(296): Set testtb-testExportWithChecksum to state=DISABLED 2025-01-08T02:23:44,964 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=191, state=SUCCESS; DisableTableProcedure table=testtb-testExportWithChecksum in 176 msec 2025-01-08T02:23:45,091 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=191 2025-01-08T02:23:45,091 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportWithChecksum, procId: 191 completed 2025-01-08T02:23:45,091 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportWithChecksum 2025-01-08T02:23:45,092 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=197, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:23:45,093 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=197, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:23:45,093 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportWithChecksum 2025-01-08T02:23:45,093 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=197, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:23:45,094 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportWithChecksum 2025-01-08T02:23:45,096 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:23:45,096 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e 2025-01-08T02:23:45,097 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,097 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,097 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,097 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,098 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T02:23:45,098 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T02:23:45,098 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T02:23:45,098 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportWithChecksum with data PBUF 2025-01-08T02:23:45,098 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/recovered.edits] 2025-01-08T02:23:45,099 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/recovered.edits] 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportWithChecksum 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,100 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,101 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=197 2025-01-08T02:23:45,101 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,101 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,101 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,102 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,103 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/cf/57d3f708e3094de081f1750a305aa4b4 2025-01-08T02:23:45,104 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/cf/a42ae286b9d2472ba3ee13971ce93f7d to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/cf/a42ae286b9d2472ba3ee13971ce93f7d 2025-01-08T02:23:45,106 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e/recovered.edits/9.seqid 2025-01-08T02:23:45,106 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1/recovered.edits/9.seqid 2025-01-08T02:23:45,107 DEBUG [HFileArchiver-20 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/a88b28b80a439cdaac59080bea96479e 2025-01-08T02:23:45,107 DEBUG [HFileArchiver-19 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportWithChecksum/309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:23:45,107 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportWithChecksum regions 2025-01-08T02:23:45,107 DEBUG [PEWorker-5 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6 2025-01-08T02:23:45,108 DEBUG [PEWorker-5 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf] 2025-01-08T02:23:45,110 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b2025010893a73f1f282844c588aaca8218502257_a88b28b80a439cdaac59080bea96479e to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/c4ca4238a0b923820dcc509a6f75849b2025010893a73f1f282844c588aaca8218502257_a88b28b80a439cdaac59080bea96479e 2025-01-08T02:23:45,111 DEBUG [PEWorker-5 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108882316213c98464bbd4e216579f0fa5f_309344e6ac66af78cd30d2baec21e5b1 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6/cf/d41d8cd98f00b204e9800998ecf8427e20250108882316213c98464bbd4e216579f0fa5f_309344e6ac66af78cd30d2baec21e5b1 2025-01-08T02:23:45,112 DEBUG [PEWorker-5 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportWithChecksum/079394da860334b7f5313f35a50f5bc6 2025-01-08T02:23:45,113 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=197, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:23:45,115 WARN [PEWorker-5 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportWithChecksum from hbase:meta 2025-01-08T02:23:45,117 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportWithChecksum' descriptor. 2025-01-08T02:23:45,117 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=197, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:23:45,117 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportWithChecksum' from region states. 2025-01-08T02:23:45,118 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736303025117"}]},"ts":"9223372036854775807"} 2025-01-08T02:23:45,118 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736303025117"}]},"ts":"9223372036854775807"} 2025-01-08T02:23:45,119 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:23:45,119 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 309344e6ac66af78cd30d2baec21e5b1, NAME => 'testtb-testExportWithChecksum,,1736302974660.309344e6ac66af78cd30d2baec21e5b1.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => a88b28b80a439cdaac59080bea96479e, NAME => 'testtb-testExportWithChecksum,1,1736302974660.a88b28b80a439cdaac59080bea96479e.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:23:45,119 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportWithChecksum' as deleted. 2025-01-08T02:23:45,119 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportWithChecksum","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736303025119"}]},"ts":"9223372036854775807"} 2025-01-08T02:23:45,121 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportWithChecksum state from META 2025-01-08T02:23:45,123 DEBUG [PEWorker-5 {}] procedure.DeleteTableProcedure(133): Finished pid=197, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportWithChecksum 2025-01-08T02:23:45,123 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=197, state=SUCCESS; DeleteTableProcedure table=testtb-testExportWithChecksum in 32 msec 2025-01-08T02:23:45,202 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=197 2025-01-08T02:23:45,202 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportWithChecksum, procId: 197 completed 2025-01-08T02:23:45,207 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportWithChecksum" 2025-01-08T02:23:45,208 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportWithChecksum 2025-01-08T02:23:45,209 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportWithChecksum" 2025-01-08T02:23:45,210 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportWithChecksum 2025-01-08T02:23:45,234 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportWithChecksum Thread=802 (was 809), OpenFileDescriptor=813 (was 821), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=464 (was 560), ProcessCount=18 (was 15) - ProcessCount LEAK? -, AvailableMemoryMB=779 (was 2152) 2025-01-08T02:23:45,234 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=802 is superior to 500 2025-01-08T02:23:45,251 INFO [Time-limited test {}] hbase.ResourceChecker(147): before: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithSkipTmp Thread=802, OpenFileDescriptor=813, MaxFileDescriptor=1048576, SystemLoadAverage=464, ProcessCount=18, AvailableMemoryMB=784 2025-01-08T02:23:45,251 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=802 is superior to 500 2025-01-08T02:23:45,252 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$4(2404): Client=jenkins//172.17.0.3 create 'testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'} 2025-01-08T02:23:45,253 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=198, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:45,253 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_PRE_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_PRE_OPERATION 2025-01-08T02:23:45,253 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(713): Client=jenkins//172.17.0.3 procedure request for creating table: namespace: "default" qualifier: "testtb-testExportFileSystemStateWithSkipTmp" procId is: 198 2025-01-08T02:23:45,254 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_WRITE_FS_LAYOUT, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_WRITE_FS_LAYOUT 2025-01-08T02:23:45,254 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T02:23:45,259 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742372_1548 (size=454) 2025-01-08T02:23:45,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742372_1548 (size=454) 2025-01-08T02:23:45,260 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742372_1548 (size=454) 2025-01-08T02:23:45,261 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(7106): creating {ENCODED => 947fe374ffc550ab7ab42fb17014d00e, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.', STARTKEY => '', ENDKEY => '1'}, tableDescriptor='testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:23:45,262 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(7106): creating {ENCODED => 85b534583da52ed359fa76f762619b11, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.', STARTKEY => '1', ENDKEY => ''}, tableDescriptor='testtb-testExportFileSystemStateWithSkipTmp', {TABLE_ATTRIBUTES => {REGION_REPLICATION => '1', METADATA => {'hbase.store.file-tracker.impl' => 'DEFAULT'}}}, {NAME => 'cf', INDEX_BLOCK_ENCODING => 'NONE', MOB_THRESHOLD => '0', VERSIONS => '1', KEEP_DELETED_CELLS => 'FALSE', DATA_BLOCK_ENCODING => 'NONE', TTL => 'FOREVER', MIN_VERSIONS => '0', REPLICATION_SCOPE => '0', BLOOMFILTER => 'ROW', IN_MEMORY => 'false', IS_MOB => 'true', COMPRESSION => 'NONE', BLOCKCACHE => 'true', BLOCKSIZE => '65536 B (64KB)'}, regionDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:23:45,267 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742373_1549 (size=79) 2025-01-08T02:23:45,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742374_1550 (size=79) 2025-01-08T02:23:45,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742373_1549 (size=79) 2025-01-08T02:23:45,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742373_1549 (size=79) 2025-01-08T02:23:45,268 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742374_1550 (size=79) 2025-01-08T02:23:45,269 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742374_1550 (size=79) 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1681): Closing 947fe374ffc550ab7ab42fb17014d00e, disabling compactions & flushes 2025-01-08T02:23:45,269 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. after waiting 0 ms 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,269 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-0 {}] regionserver.HRegion(1635): Region close journal for 947fe374ffc550ab7ab42fb17014d00e: 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1681): Closing 85b534583da52ed359fa76f762619b11, disabling compactions & flushes 2025-01-08T02:23:45,269 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. after waiting 0 ms 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:45,269 INFO [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:45,269 DEBUG [RegionOpenAndInit-testtb-testExportFileSystemStateWithSkipTmp-pool-1 {}] regionserver.HRegion(1635): Region close journal for 85b534583da52ed359fa76f762619b11: 2025-01-08T02:23:45,270 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_ADD_TO_META, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_ADD_TO_META 2025-01-08T02:23:45,270 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.","families":{"info":[{"qualifier":"regioninfo","vlen":78,"tag":[],"timestamp":"1736303025270"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736303025270"}]},"ts":"1736303025270"} 2025-01-08T02:23:45,271 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":2,"row":"testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.","families":{"info":[{"qualifier":"regioninfo","vlen":78,"tag":[],"timestamp":"1736303025270"},{"qualifier":"state","vlen":6,"tag":[],"timestamp":"1736303025270"}]},"ts":"1736303025270"} 2025-01-08T02:23:45,272 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1516): Added 2 regions to meta. 2025-01-08T02:23:45,273 INFO [PEWorker-4 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_ASSIGN_REGIONS, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_ASSIGN_REGIONS 2025-01-08T02:23:45,273 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736303025273"}]},"ts":"1736303025273"} 2025-01-08T02:23:45,274 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=ENABLING in hbase:meta 2025-01-08T02:23:45,278 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(202): Hosts are {c4e5d52b9759=0} racks are {/default-rack=0} 2025-01-08T02:23:45,279 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 0 is on host 0 2025-01-08T02:23:45,279 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 1 is on host 0 2025-01-08T02:23:45,279 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(303): server 2 is on host 0 2025-01-08T02:23:45,279 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 0 is on rack 0 2025-01-08T02:23:45,279 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 1 is on rack 0 2025-01-08T02:23:45,279 INFO [PEWorker-4 {}] balancer.BalancerClusterState(314): server 2 is on rack 0 2025-01-08T02:23:45,279 DEBUG [PEWorker-4 {}] balancer.BalancerClusterState(319): Number of tables=1, number of hosts=1, number of racks=1 2025-01-08T02:23:45,279 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=947fe374ffc550ab7ab42fb17014d00e, ASSIGN}, {pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=85b534583da52ed359fa76f762619b11, ASSIGN}] 2025-01-08T02:23:45,280 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=85b534583da52ed359fa76f762619b11, ASSIGN 2025-01-08T02:23:45,280 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=947fe374ffc550ab7ab42fb17014d00e, ASSIGN 2025-01-08T02:23:45,280 INFO [PEWorker-1 {}] assignment.TransitRegionStateProcedure(264): Starting pid=200, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=85b534583da52ed359fa76f762619b11, ASSIGN; state=OFFLINE, location=c4e5d52b9759,39911,1736302808913; forceNewPlan=false, retain=false 2025-01-08T02:23:45,280 INFO [PEWorker-3 {}] assignment.TransitRegionStateProcedure(264): Starting pid=199, ppid=198, state=RUNNABLE:REGION_STATE_TRANSITION_GET_ASSIGN_CANDIDATE, locked=true; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=947fe374ffc550ab7ab42fb17014d00e, ASSIGN; state=OFFLINE, location=c4e5d52b9759,36743,1736302808837; forceNewPlan=false, retain=false 2025-01-08T02:23:45,355 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T02:23:45,431 INFO [c4e5d52b9759:41063 {}] balancer.BaseLoadBalancer(546): Reassigned 2 regions. 2 retained the pre-restart assignment. 2025-01-08T02:23:45,431 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=200 updating hbase:meta row=85b534583da52ed359fa76f762619b11, regionState=OPENING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:23:45,431 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=199 updating hbase:meta row=947fe374ffc550ab7ab42fb17014d00e, regionState=OPENING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:23:45,433 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=201, ppid=199, state=RUNNABLE; OpenRegionProcedure 947fe374ffc550ab7ab42fb17014d00e, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:23:45,433 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=202, ppid=200, state=RUNNABLE; OpenRegionProcedure 85b534583da52ed359fa76f762619b11, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:23:45,555 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T02:23:45,584 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:23:45,585 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:23:45,587 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,587 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7285): Opening region: {ENCODED => 947fe374ffc550ab7ab42fb17014d00e, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.', STARTKEY => '', ENDKEY => '1'} 2025-01-08T02:23:45,587 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] handler.AssignRegionHandler(135): Open testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:45,587 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. service=AccessControlService 2025-01-08T02:23:45,587 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7285): Opening region: {ENCODED => 85b534583da52ed359fa76f762619b11, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.', STARTKEY => '1', ENDKEY => ''} 2025-01-08T02:23:45,587 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7999): Registered coprocessor service: region=testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. service=AccessControlService 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithSkipTmp 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:23:45,588 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] coprocessor.CoprocessorHost(174): System coprocessor org.apache.hadoop.hbase.security.access.AccessController loaded, priority=536870911. 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7327): checking encryption for 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(7330): checking classloading for 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.MetricsRegionSourceImpl(79): Creating new MetricsRegionSourceImpl for table testtb-testExportFileSystemStateWithSkipTmp 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(894): Instantiated testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.; StoreHotnessProtector, parallelPutToStoreThreadLimit=0 ; minColumnNum=100 ; preparePutThreadLimit=0 ; hotProtect now disable 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7327): checking encryption for 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,588 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(7330): checking classloading for 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,589 INFO [StoreOpener-947fe374ffc550ab7ab42fb17014d00e-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,589 INFO [StoreOpener-85b534583da52ed359fa76f762619b11-1 {}] regionserver.HStore(400): Created cacheConfig: cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false, for column family cf of region 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,590 INFO [StoreOpener-947fe374ffc550ab7ab42fb17014d00e-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 947fe374ffc550ab7ab42fb17014d00e columnFamilyName cf 2025-01-08T02:23:45,590 INFO [StoreOpener-85b534583da52ed359fa76f762619b11-1 {}] compactions.CompactionConfiguration(181): size [minCompactSize:128 MB, maxCompactSize:8.00 EB, offPeakMaxCompactSize:8.00 EB); files [minFilesToCompact:3, maxFilesToCompact:10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000; tiered compaction: max_age 9223372036854775807, incoming window min 6, compaction policy for tiered window org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy, single output for minor true, compaction window factory org.apache.hadoop.hbase.regionserver.compactions.ExponentialCompactionWindowFactory, region 85b534583da52ed359fa76f762619b11 columnFamilyName cf 2025-01-08T02:23:45,591 DEBUG [StoreOpener-947fe374ffc550ab7ab42fb17014d00e-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:23:45,591 DEBUG [StoreOpener-85b534583da52ed359fa76f762619b11-1 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:23:45,591 INFO [StoreOpener-947fe374ffc550ab7ab42fb17014d00e-1 {}] regionserver.HStore(327): Store=947fe374ffc550ab7ab42fb17014d00e/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:23:45,591 INFO [StoreOpener-85b534583da52ed359fa76f762619b11-1 {}] regionserver.HStore(327): Store=85b534583da52ed359fa76f762619b11/cf, memstore type=DefaultMemStore, storagePolicy=NONE, verifyBulkLoads=false, parallelPutCountPrintThreshold=50, encoding=NONE, compression=NONE 2025-01-08T02:23:45,592 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,592 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,592 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,592 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(5301): Found 0 recovered edits file(s) under hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,594 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1085): writing seq id for 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,594 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1085): writing seq id for 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,595 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:23:45,595 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/recovered.edits/1.seqid, newMaxSeqId=1, maxSeqId=-1 2025-01-08T02:23:45,596 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1102): Opened 947fe374ffc550ab7ab42fb17014d00e; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=71451722, jitterRate=0.06471362709999084}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:23:45,596 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1102): Opened 85b534583da52ed359fa76f762619b11; next sequenceid=2; SteppingSplitPolicysuper{IncreasingToUpperBoundRegionSplitPolicy{initialSize=268435456, ConstantSizeRegionSplitPolicy{desiredMaxFileSize=74861231, jitterRate=0.11551927030086517}}}, FlushLargeStoresPolicy{flushSizeLowerBound=-1} 2025-01-08T02:23:45,596 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegion(1001): Region open journal for 947fe374ffc550ab7ab42fb17014d00e: 2025-01-08T02:23:45,596 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegion(1001): Region open journal for 85b534583da52ed359fa76f762619b11: 2025-01-08T02:23:45,597 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11., pid=202, masterSystemTime=1736303025585 2025-01-08T02:23:45,597 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegionServer(2601): Post open deploy tasks for testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e., pid=201, masterSystemTime=1736303025584 2025-01-08T02:23:45,598 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,598 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=201}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,599 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=199 updating hbase:meta row=947fe374ffc550ab7ab42fb17014d00e, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:23:45,599 DEBUG [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] regionserver.HRegionServer(2628): Finished post open deploy task for testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:45,599 INFO [RS_OPEN_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_OPEN_REGION, pid=202}] handler.AssignRegionHandler(164): Opened testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:45,599 INFO [PEWorker-5 {}] assignment.RegionStateStore(202): pid=200 updating hbase:meta row=85b534583da52ed359fa76f762619b11, regionState=OPEN, openSeqNum=2, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:23:45,601 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=201, resume processing ppid=199 2025-01-08T02:23:45,601 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=201, ppid=199, state=SUCCESS; OpenRegionProcedure 947fe374ffc550ab7ab42fb17014d00e, server=c4e5d52b9759,36743,1736302808837 in 168 msec 2025-01-08T02:23:45,602 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=202, resume processing ppid=200 2025-01-08T02:23:45,602 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=202, ppid=200, state=SUCCESS; OpenRegionProcedure 85b534583da52ed359fa76f762619b11, server=c4e5d52b9759,39911,1736302808913 in 168 msec 2025-01-08T02:23:45,603 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=199, ppid=198, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=947fe374ffc550ab7ab42fb17014d00e, ASSIGN in 322 msec 2025-01-08T02:23:45,604 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=200, resume processing ppid=198 2025-01-08T02:23:45,604 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=200, ppid=198, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=85b534583da52ed359fa76f762619b11, ASSIGN in 323 msec 2025-01-08T02:23:45,604 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_UPDATE_DESC_CACHE, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_UPDATE_DESC_CACHE 2025-01-08T02:23:45,605 DEBUG [PEWorker-3 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736303025604"}]},"ts":"1736303025604"} 2025-01-08T02:23:45,605 INFO [PEWorker-3 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=ENABLED in hbase:meta 2025-01-08T02:23:45,608 INFO [PEWorker-3 {}] procedure.CreateTableProcedure(89): pid=198, state=RUNNABLE:CREATE_TABLE_POST_OPERATION, locked=true; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp execute state=CREATE_TABLE_POST_OPERATION 2025-01-08T02:23:45,608 DEBUG [PEWorker-3 {}] access.PermissionStorage(175): Writing permission with rowKey testtb-testExportFileSystemStateWithSkipTmp jenkins: RWXCA 2025-01-08T02:23:45,609 DEBUG [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T02:23:45,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,614 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,616 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF\x0AQ\x0A\x07jenkins\x12F\x08\x03"B\x0A6\x0A\x07default\x12+testtb-testExportFileSystemStateWithSkipTmp \x00 \x01 \x02 \x03 \x04 2025-01-08T02:23:45,617 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=198, state=SUCCESS; CreateTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 363 msec 2025-01-08T02:23:45,856 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=198 2025-01-08T02:23:45,857 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: CREATE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 198 completed 2025-01-08T02:23:45,859 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:45,859 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:45,859 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:23:45,868 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T02:23:45,868 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736303025868 (current time:1736303025868). 2025-01-08T02:23:45,868 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:23:45,868 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp VERSION not specified, setting to 2 2025-01-08T02:23:45,868 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:23:45,869 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x4a129144 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@5d141eec 2025-01-08T02:23:45,873 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@633f2f1e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:23:45,874 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:23:45,875 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39912, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:23:45,876 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x4a129144 to 127.0.0.1:61593 2025-01-08T02:23:45,876 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:23:45,876 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x6adee2a2 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@17c0626b 2025-01-08T02:23:45,879 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@d1df852, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:23:45,881 DEBUG [hconnection-0x7f19f2d8-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:23:45,882 INFO [RS-EventLoopGroup-4-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39920, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:23:45,883 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:23:45,884 INFO [RS-EventLoopGroup-3-1 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35920, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:23:45,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x6adee2a2 to 127.0.0.1:61593 2025-01-08T02:23:45,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:23:45,885 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T02:23:45,886 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:23:45,886 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=203, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T02:23:45,887 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 203 2025-01-08T02:23:45,887 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:23:45,887 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T02:23:45,888 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:23:45,890 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:23:45,894 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742375_1551 (size=203) 2025-01-08T02:23:45,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742375_1551 (size=203) 2025-01-08T02:23:45,895 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742375_1551 (size=203) 2025-01-08T02:23:45,896 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:23:45,896 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e}, {pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11}] 2025-01-08T02:23:45,897 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:45,897 INFO [PEWorker-2 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:45,988 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T02:23:46,048 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:23:46,048 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:23:46,048 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=204 2025-01-08T02:23:46,048 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=205 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.HRegion(2538): Flush status journal for 947fe374ffc550ab7ab42fb17014d00e: 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.HRegion(2538): Flush status journal for 85b534583da52ed359fa76f762619b11: 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. for emptySnaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. for emptySnaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.' region-info for snapshot=emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:23:46,049 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] snapshot.SnapshotManifest(256): Adding snapshot references for [] hfiles 2025-01-08T02:23:46,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742376_1552 (size=82) 2025-01-08T02:23:46,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742377_1553 (size=82) 2025-01-08T02:23:46,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742377_1553 (size=82) 2025-01-08T02:23:46,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742377_1553 (size=82) 2025-01-08T02:23:46,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742376_1552 (size=82) 2025-01-08T02:23:46,056 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:46,056 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=204}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=204 2025-01-08T02:23:46,056 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742376_1552 (size=82) 2025-01-08T02:23:46,056 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=204 2025-01-08T02:23:46,056 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp on region 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:46,056 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:46,056 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-1 {event_type=RS_SNAPSHOT_REGIONS, pid=205}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=205 2025-01-08T02:23:46,056 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=204, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:46,057 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=205 2025-01-08T02:23:46,057 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot emptySnaptb0-testExportFileSystemStateWithSkipTmp on region 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:46,057 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=205, ppid=203, state=RUNNABLE; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:46,058 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=204, ppid=203, state=SUCCESS; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e in 161 msec 2025-01-08T02:23:46,059 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=205, resume processing ppid=203 2025-01-08T02:23:46,059 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=205, ppid=203, state=SUCCESS; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11 in 161 msec 2025-01-08T02:23:46,059 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:23:46,059 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:23:46,060 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:23:46,060 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:23:46,060 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:23:46,060 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(328): No files under family: cf 2025-01-08T02:23:46,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742378_1554 (size=74) 2025-01-08T02:23:46,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742378_1554 (size=74) 2025-01-08T02:23:46,065 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742378_1554 (size=74) 2025-01-08T02:23:46,066 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:23:46,066 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,066 DEBUG [PEWorker-1 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,072 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742379_1555 (size=697) 2025-01-08T02:23:46,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742379_1555 (size=697) 2025-01-08T02:23:46,073 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742379_1555 (size=697) 2025-01-08T02:23:46,075 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:23:46,078 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:23:46,078 DEBUG [PEWorker-1 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/emptySnaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,079 INFO [PEWorker-1 {}] procedure.SnapshotProcedure(134): pid=203, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:23:46,079 DEBUG [PEWorker-1 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 203 2025-01-08T02:23:46,080 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=203, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=203, snapshot={ ss=emptySnaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } in 193 msec 2025-01-08T02:23:46,189 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=203 2025-01-08T02:23:46,189 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 203 completed 2025-01-08T02:23:46,191 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=36743 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:23:46,192 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=39911 {}] regionserver.HRegion(8254): writing data to region testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. with WAL disabled. Data may be lost in the event of a crash. 2025-01-08T02:23:46,195 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2718): Found 2 regions for table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,195 DEBUG [Time-limited test {}] hbase.HBaseTestingUtility(2724): firstRegionName=testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:46,195 INFO [Time-limited test {}] hbase.Waiter(181): Waiting up to [60,000] milli-secs(wait.for.ratio=[1]) 2025-01-08T02:23:46,206 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1703): Client=jenkins//172.17.0.3 snapshot request for:{ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T02:23:46,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(331): Creation time not specified, setting to:1736303026206 (current time:1736303026206). 2025-01-08T02:23:46,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(345): Snapshot current TTL value: 0 resetting it to default value: 0 2025-01-08T02:23:46,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(354): Snapshot snaptb0-testExportFileSystemStateWithSkipTmp VERSION not specified, setting to 2 2025-01-08T02:23:46,206 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotDescriptionUtils(361): Set jenkins as owner of Snapshot 2025-01-08T02:23:46,207 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x387a4382 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@f4e6ed7 2025-01-08T02:23:46,218 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@378756fe, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:23:46,221 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:23:46,222 INFO [RS-EventLoopGroup-4-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39936, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:23:46,223 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x387a4382 to 127.0.0.1:61593 2025-01-08T02:23:46,223 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:23:46,224 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(149): Connect 0x28cbcf64 to 127.0.0.1:61593 with session timeout=90000ms, retries=30, retry interval=1000ms, keepAlive=60000ms, zk client config=org.apache.zookeeper.client.ZKClientConfig@540f347c 2025-01-08T02:23:46,232 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@5932a8a5, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=null 2025-01-08T02:23:46,234 DEBUG [hconnection-0x241ff9e4-metaLookup-shared--pool-0 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:23:46,235 INFO [RS-EventLoopGroup-4-3 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:39946, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:23:46,236 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.RpcConnection(159): Using SIMPLE authentication for service=ClientService, sasl=false 2025-01-08T02:23:46,237 INFO [RS-EventLoopGroup-3-2 {}] ipc.ServerRpcConnection(484): Connection from 172.17.0.3:35930, version=2.7.0-SNAPSHOT, sasl=false, ugi=jenkins (auth:SIMPLE), service=ClientService 2025-01-08T02:23:46,238 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x28cbcf64 to 127.0.0.1:61593 2025-01-08T02:23:46,238 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:23:46,238 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(611): Read acl: entry[testtb-testExportFileSystemStateWithSkipTmp], kv [jenkins: RWXCA] 2025-01-08T02:23:46,239 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(806): No existing snapshot, attempting snapshot... 2025-01-08T02:23:46,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=206, state=RUNNABLE:SNAPSHOT_PREPARE; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } 2025-01-08T02:23:46,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(1441): register snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 206 2025-01-08T02:23:46,240 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T02:23:46,240 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_PREPARE, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PREPARE 2025-01-08T02:23:46,241 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_PRE_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_PRE_OPERATION 2025-01-08T02:23:46,243 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_WRITE_SNAPSHOT_INFO, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_WRITE_SNAPSHOT_INFO 2025-01-08T02:23:46,252 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742380_1556 (size=198) 2025-01-08T02:23:46,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742380_1556 (size=198) 2025-01-08T02:23:46,253 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742380_1556 (size=198) 2025-01-08T02:23:46,254 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_ONLINE_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_ONLINE_REGIONS 2025-01-08T02:23:46,254 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e}, {pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11}] 2025-01-08T02:23:46,255 INFO [PEWorker-5 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:46,255 INFO [PEWorker-4 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:46,341 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T02:23:46,405 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:23:46,406 DEBUG [RSProcedureDispatcher-pool-2 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:23:46,406 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=39911 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=208 2025-01-08T02:23:46,406 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:46,406 DEBUG [RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=36743 {}] regionserver.RSRpcServices(3992): Executing remote procedure class org.apache.hadoop.hbase.regionserver.SnapshotRegionCallable, pid=207 2025-01-08T02:23:46,406 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(50): Starting snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:46,407 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(2837): Flushing 85b534583da52ed359fa76f762619b11 1/1 column families, dataSize=3.00 KB heapSize=6.72 KB 2025-01-08T02:23:46,407 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(2837): Flushing 947fe374ffc550ab7ab42fb17014d00e 1/1 column families, dataSize=266 B heapSize=832 B 2025-01-08T02:23:46,425 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108da8c61020b7b45279d4a0198c20f55f4_85b534583da52ed359fa76f762619b11 is 71, key is 1495139c6e60a88d531506a1c665ac18/cf:q/1736303026192/Put/seqid=0 2025-01-08T02:23:46,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742381_1557 (size=8101) 2025-01-08T02:23:46,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742381_1557 (size=8101) 2025-01-08T02:23:46,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742381_1557 (size=8101) 2025-01-08T02:23:46,430 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108d27cad870ca5438fae93a54470fb486f_947fe374ffc550ab7ab42fb17014d00e is 71, key is 00698e5d5190b81a3915c200024c5216/cf:q/1736303026191/Put/seqid=0 2025-01-08T02:23:46,430 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:23:46,434 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/c4ca4238a0b923820dcc509a6f75849b20250108da8c61020b7b45279d4a0198c20f55f4_85b534583da52ed359fa76f762619b11 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108da8c61020b7b45279d4a0198c20f55f4_85b534583da52ed359fa76f762619b11 2025-01-08T02:23:46,435 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742382_1558 (size=5171) 2025-01-08T02:23:46,435 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742382_1558 (size=5171) 2025-01-08T02:23:46,435 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742382_1558 (size=5171) 2025-01-08T02:23:46,435 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/.tmp/cf/2ba935ee465643b5a27d9ad9371f2d13, store: [table=testtb-testExportFileSystemStateWithSkipTmp family=cf region=85b534583da52ed359fa76f762619b11] 2025-01-08T02:23:46,435 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:23:46,436 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/.tmp/cf/2ba935ee465643b5a27d9ad9371f2d13 is 220, key is 1ef30d071885d53bbca2e387fc9598af1/cf:q/1736303026192/Put/seqid=0 2025-01-08T02:23:46,439 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HMobStore(268): FLUSH Renaming flushed file from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/.tmp/d41d8cd98f00b204e9800998ecf8427e20250108d27cad870ca5438fae93a54470fb486f_947fe374ffc550ab7ab42fb17014d00e to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e20250108d27cad870ca5438fae93a54470fb486f_947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:46,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742383_1559 (size=15309) 2025-01-08T02:23:46,440 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742383_1559 (size=15309) 2025-01-08T02:23:46,441 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] mob.DefaultMobStoreFlusher(263): Flush store file: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/.tmp/cf/d4958f48a1a64f829999f946a2049531, store: [table=testtb-testExportFileSystemStateWithSkipTmp family=cf region=947fe374ffc550ab7ab42fb17014d00e] 2025-01-08T02:23:46,441 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742383_1559 (size=15309) 2025-01-08T02:23:46,441 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/.tmp/cf/d4958f48a1a64f829999f946a2049531 is 220, key is 0bb5a10e8e83c3d62c7b296342a945543/cf:q/1736303026191/Put/seqid=0 2025-01-08T02:23:46,442 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=3.0 K, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/.tmp/cf/2ba935ee465643b5a27d9ad9371f2d13 2025-01-08T02:23:46,445 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742384_1560 (size=6176) 2025-01-08T02:23:46,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742384_1560 (size=6176) 2025-01-08T02:23:46,446 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742384_1560 (size=6176) 2025-01-08T02:23:46,446 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/.tmp/cf/2ba935ee465643b5a27d9ad9371f2d13 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/cf/2ba935ee465643b5a27d9ad9371f2d13 2025-01-08T02:23:46,446 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] mob.DefaultMobStoreFlusher(147): Mob store is flushed, sequenceid=6, memsize=266, hasBloomFilter=true, into tmp file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/.tmp/cf/d4958f48a1a64f829999f946a2049531 2025-01-08T02:23:46,450 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/.tmp/cf/d4958f48a1a64f829999f946a2049531 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/cf/d4958f48a1a64f829999f946a2049531 2025-01-08T02:23:46,450 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/cf/2ba935ee465643b5a27d9ad9371f2d13, entries=46, sequenceid=6, filesize=15.0 K 2025-01-08T02:23:46,451 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(3040): Finished flush of dataSize ~3.00 KB/3070, heapSize ~6.70 KB/6864, currentSize=0 B/0 for 85b534583da52ed359fa76f762619b11 in 45ms, sequenceid=6, compaction requested=false 2025-01-08T02:23:46,451 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.MetricsTableSourceImpl(133): Creating new MetricsTableSourceImpl for table 'testtb-testExportFileSystemStateWithSkipTmp' 2025-01-08T02:23:46,451 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.HRegion(2538): Flush status journal for 85b534583da52ed359fa76f762619b11: 2025-01-08T02:23:46,451 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. for snaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T02:23:46,452 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.' region-info for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,452 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:23:46,452 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/cf/2ba935ee465643b5a27d9ad9371f2d13] hfiles 2025-01-08T02:23:46,452 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/cf/2ba935ee465643b5a27d9ad9371f2d13 for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,454 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/cf/d4958f48a1a64f829999f946a2049531, entries=4, sequenceid=6, filesize=6.0 K 2025-01-08T02:23:46,454 INFO [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(3040): Finished flush of dataSize ~266 B/266, heapSize ~816 B/816, currentSize=0 B/0 for 947fe374ffc550ab7ab42fb17014d00e in 47ms, sequenceid=6, compaction requested=false 2025-01-08T02:23:46,454 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.HRegion(2538): Flush status journal for 947fe374ffc550ab7ab42fb17014d00e: 2025-01-08T02:23:46,454 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(75): Snapshotting region testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. for snaptb0-testExportFileSystemStateWithSkipTmp completed. 2025-01-08T02:23:46,455 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(241): Storing 'testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.' region-info for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,455 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(246): Creating references for hfiles 2025-01-08T02:23:46,455 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(256): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/cf/d4958f48a1a64f829999f946a2049531] hfiles 2025-01-08T02:23:46,455 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] snapshot.SnapshotManifest(265): Adding reference for file (1/1): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/cf/d4958f48a1a64f829999f946a2049531 for snapshot=snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742385_1561 (size=121) 2025-01-08T02:23:46,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742385_1561 (size=121) 2025-01-08T02:23:46,458 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742385_1561 (size=121) 2025-01-08T02:23:46,459 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:23:46,459 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-2 {event_type=RS_SNAPSHOT_REGIONS, pid=208}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=208 2025-01-08T02:23:46,459 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=208 2025-01-08T02:23:46,459 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithSkipTmp on region 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:46,459 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=208, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11 2025-01-08T02:23:46,461 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=208, ppid=206, state=SUCCESS; SnapshotRegionProcedure 85b534583da52ed359fa76f762619b11 in 206 msec 2025-01-08T02:23:46,461 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742386_1562 (size=121) 2025-01-08T02:23:46,462 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742386_1562 (size=121) 2025-01-08T02:23:46,462 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742386_1562 (size=121) 2025-01-08T02:23:46,462 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.SnapshotRegionCallable(78): Closing snapshot operation on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:23:46,462 DEBUG [RS_SNAPSHOT_OPERATIONS-regionserver/c4e5d52b9759:0-0 {event_type=RS_SNAPSHOT_REGIONS, pid=207}] regionserver.RemoteProcedureResultReporter(62): Successfully complete execution of pid=207 2025-01-08T02:23:46,463 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster(4121): Remote procedure done, pid=207 2025-01-08T02:23:46,463 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure.SnapshotRegionProcedure(132): finish snapshot snaptb0-testExportFileSystemStateWithSkipTmp on region 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:46,463 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=207, ppid=206, state=RUNNABLE; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:46,465 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=207, resume processing ppid=206 2025-01-08T02:23:46,465 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_SPLIT_REGIONS, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_SPLIT_REGIONS 2025-01-08T02:23:46,465 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=207, ppid=206, state=SUCCESS; SnapshotRegionProcedure 947fe374ffc550ab7ab42fb17014d00e in 209 msec 2025-01-08T02:23:46,465 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_SNAPSHOT_MOB_REGION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_SNAPSHOT_MOB_REGION 2025-01-08T02:23:46,466 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(300): Storing region-info for snapshot. 2025-01-08T02:23:46,466 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(305): Creating references for hfiles 2025-01-08T02:23:46,466 DEBUG [MobRegionSnapshotPool-pool-0 {}] storefiletracker.StoreFileTrackerFactory(122): instantiating StoreFileTracker impl org.apache.hadoop.hbase.regionserver.storefiletracker.DefaultStoreFileTracker 2025-01-08T02:23:46,467 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(366): Adding snapshot references for [hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108da8c61020b7b45279d4a0198c20f55f4_85b534583da52ed359fa76f762619b11, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e20250108d27cad870ca5438fae93a54470fb486f_947fe374ffc550ab7ab42fb17014d00e] hfiles 2025-01-08T02:23:46,467 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (1/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108da8c61020b7b45279d4a0198c20f55f4_85b534583da52ed359fa76f762619b11 2025-01-08T02:23:46,467 DEBUG [MobRegionSnapshotPool-pool-0 {}] snapshot.SnapshotManifest(374): Adding reference for hfile (2/2): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e20250108d27cad870ca5438fae93a54470fb486f_947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:23:46,471 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742387_1563 (size=305) 2025-01-08T02:23:46,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742387_1563 (size=305) 2025-01-08T02:23:46,472 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742387_1563 (size=305) 2025-01-08T02:23:46,473 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_CONSOLIDATE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_CONSOLIDATE_SNAPSHOT 2025-01-08T02:23:46,473 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifest(495): Convert to Single Snapshot Manifest for snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,473 DEBUG [PEWorker-2 {}] snapshot.SnapshotManifestV1(130): No regions under directory:hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,479 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742388_1564 (size=1007) 2025-01-08T02:23:46,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742388_1564 (size=1007) 2025-01-08T02:23:46,480 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742388_1564 (size=1007) 2025-01-08T02:23:46,482 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_VERIFIER_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_VERIFIER_SNAPSHOT 2025-01-08T02:23:46,486 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_COMPLETE_SNAPSHOT, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_COMPLETE_SNAPSHOT 2025-01-08T02:23:46,486 DEBUG [PEWorker-2 {}] snapshot.SnapshotDescriptionUtils(430): Sentinel is done, just moving the snapshot from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/.tmp/snaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,487 INFO [PEWorker-2 {}] procedure.SnapshotProcedure(134): pid=206, state=RUNNABLE:SNAPSHOT_POST_OPERATION, locked=true; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } execute state=SNAPSHOT_POST_OPERATION 2025-01-08T02:23:46,487 DEBUG [PEWorker-2 {}] snapshot.SnapshotManager(1447): unregister snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 }, snapshot procedure id = 206 2025-01-08T02:23:46,488 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=206, state=SUCCESS; org.apache.hadoop.hbase.master.procedure.SnapshotProcedure, id=206, snapshot={ ss=snaptb0-testExportFileSystemStateWithSkipTmp table=testtb-testExportFileSystemStateWithSkipTmp type=FLUSH ttl=0 } in 248 msec 2025-01-08T02:23:46,542 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=206 2025-01-08T02:23:46,542 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: SNAPSHOT, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 206 completed 2025-01-08T02:23:46,542 INFO [Time-limited test {}] snapshot.TestExportSnapshot(468): HDFS export destination path: hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542 2025-01-08T02:23:46,542 INFO [Time-limited test {}] snapshot.TestExportSnapshot(495): tgtFsUri=hdfs://localhost:43527, tgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542, rawTgtDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542, srcFsUri=hdfs://localhost:43527, srcDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:23:46,570 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1082): inputFs=hdfs://localhost:43527, inputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572 2025-01-08T02:23:46,571 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(1083): outputFs=DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]], outputRoot=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542, skipTmp=true, initialOutputSnapshotDir=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,572 INFO [Time-limited test {}] snapshot.ExportSnapshot(1092): Verify the source snapshot's expiration status and integrity. 2025-01-08T02:23:46,576 INFO [Time-limited test {}] snapshot.ExportSnapshot(1150): Copy Snapshot Manifest from hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:46,589 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742389_1565 (size=198) 2025-01-08T02:23:46,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742389_1565 (size=198) 2025-01-08T02:23:46,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742389_1565 (size=198) 2025-01-08T02:23:46,590 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742390_1566 (size=1007) 2025-01-08T02:23:46,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742390_1566 (size=1007) 2025-01-08T02:23:46,591 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742390_1566 (size=1007) 2025-01-08T02:23:46,592 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.HConstants, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-common/target/hbase-common-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:46,592 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol/target/hbase-protocol-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:46,593 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-protocol-shaded/target/hbase-protocol-shaded-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:46,593 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.client.Put, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-client/target/hbase-client-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,653 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.ipc.RpcServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-7451812014141667722.jar 2025-01-08T02:23:47,653 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.CompatibilityFactory, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop-compat/target/hbase-hadoop-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,653 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.JobUtil, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-hadoop2-compat/target/hbase-hadoop2-compat-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,722 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.mapreduce.TableMapper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop-2860565757005669992.jar 2025-01-08T02:23:47,723 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.impl.FastLongHistogram, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics/target/hbase-metrics-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,723 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.metrics.Snapshot, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-metrics-api/target/hbase-metrics-api-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,723 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.replication.ReplicationUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-replication/target/hbase-replication-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,723 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.http.HttpServer, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-http/target/hbase-http-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,723 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.procedure2.Procedure, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-procedure/target/hbase-procedure-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,724 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.zookeeper.ZKWatcher, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-zookeeper/target/hbase-zookeeper-2.7.0-SNAPSHOT.jar 2025-01-08T02:23:47,724 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.common.collect.Lists, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-miscellaneous/4.1.9/hbase-shaded-miscellaneous-4.1.9.jar 2025-01-08T02:23:47,724 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.gson.GsonBuilder, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-gson/4.1.9/hbase-shaded-gson-4.1.9.jar 2025-01-08T02:23:47,724 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.com.google.protobuf.UnsafeByteOperations, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-protobuf/4.1.9/hbase-shaded-protobuf-4.1.9.jar 2025-01-08T02:23:47,724 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hbase.thirdparty.io.netty.channel.Channel, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-shaded-netty/4.1.9/hbase-shaded-netty-4.1.9.jar 2025-01-08T02:23:47,725 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.hbase.unsafe.HBasePlatformDependent, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hbase/thirdparty/hbase-unsafe/4.1.9/hbase-unsafe-4.1.9.jar 2025-01-08T02:23:47,725 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.zookeeper.ZooKeeper, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/zookeeper/zookeeper/3.8.4/zookeeper-3.8.4.jar 2025-01-08T02:23:47,725 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.google.protobuf.Message, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar 2025-01-08T02:23:47,725 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class com.codahale.metrics.MetricRegistry, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/dropwizard/metrics/metrics-core/3.2.6/metrics-core-3.2.6.jar 2025-01-08T02:23:47,725 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.commons.lang3.ArrayUtils, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/commons/commons-lang3/3.9/commons-lang3-3.9.jar 2025-01-08T02:23:47,725 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.api.trace.Span, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-api/1.15.0/opentelemetry-api-1.15.0.jar 2025-01-08T02:23:47,726 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.semconv.trace.attributes.SemanticAttributes, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-semconv/1.15.0-alpha/opentelemetry-semconv-1.15.0-alpha.jar 2025-01-08T02:23:47,726 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class io.opentelemetry.context.Context, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/io/opentelemetry/opentelemetry-context/1.15.0/opentelemetry-context-1.15.0.jar 2025-01-08T02:23:47,726 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:47,726 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:47,726 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.input.TextInputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:23:47,727 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.LongWritable, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:47,727 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.io.Text, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-common/3.4.1/hadoop-common-3.4.1.jar 2025-01-08T02:23:47,727 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:23:47,727 DEBUG [Time-limited test {}] mapreduce.TableMapReduceUtil(923): For class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner, using jar /home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-mapreduce-client-core/3.4.1/hadoop-mapreduce-client-core-3.4.1.jar 2025-01-08T02:23:47,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742391_1567 (size=127628) 2025-01-08T02:23:47,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742391_1567 (size=127628) 2025-01-08T02:23:47,778 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742391_1567 (size=127628) 2025-01-08T02:23:47,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T02:23:47,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T02:23:47,790 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742392_1568 (size=2172137) 2025-01-08T02:23:47,796 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742393_1569 (size=213228) 2025-01-08T02:23:47,796 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742393_1569 (size=213228) 2025-01-08T02:23:47,797 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742393_1569 (size=213228) 2025-01-08T02:23:47,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T02:23:47,806 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T02:23:47,807 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742394_1570 (size=1877034) 2025-01-08T02:23:47,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742395_1571 (size=533455) 2025-01-08T02:23:47,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742395_1571 (size=533455) 2025-01-08T02:23:47,814 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742395_1571 (size=533455) 2025-01-08T02:23:47,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T02:23:47,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T02:23:47,837 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742396_1572 (size=7280644) 2025-01-08T02:23:47,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742397_1573 (size=6350715) 2025-01-08T02:23:47,865 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742397_1573 (size=6350715) 2025-01-08T02:23:47,866 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742397_1573 (size=6350715) 2025-01-08T02:23:47,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742398_1574 (size=4188619) 2025-01-08T02:23:47,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742398_1574 (size=4188619) 2025-01-08T02:23:47,883 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742398_1574 (size=4188619) 2025-01-08T02:23:47,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742399_1575 (size=20406) 2025-01-08T02:23:47,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742399_1575 (size=20406) 2025-01-08T02:23:47,890 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742399_1575 (size=20406) 2025-01-08T02:23:47,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742400_1576 (size=75495) 2025-01-08T02:23:47,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742400_1576 (size=75495) 2025-01-08T02:23:47,897 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742400_1576 (size=75495) 2025-01-08T02:23:47,905 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742401_1577 (size=45609) 2025-01-08T02:23:47,905 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742401_1577 (size=45609) 2025-01-08T02:23:47,905 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742401_1577 (size=45609) 2025-01-08T02:23:48,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742402_1578 (size=110084) 2025-01-08T02:23:48,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742402_1578 (size=110084) 2025-01-08T02:23:48,313 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742402_1578 (size=110084) 2025-01-08T02:23:48,322 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742403_1579 (size=1323991) 2025-01-08T02:23:48,322 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742403_1579 (size=1323991) 2025-01-08T02:23:48,323 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742403_1579 (size=1323991) 2025-01-08T02:23:48,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742404_1580 (size=23076) 2025-01-08T02:23:48,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742404_1580 (size=23076) 2025-01-08T02:23:48,329 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742404_1580 (size=23076) 2025-01-08T02:23:48,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742405_1581 (size=126803) 2025-01-08T02:23:48,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742405_1581 (size=126803) 2025-01-08T02:23:48,336 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742405_1581 (size=126803) 2025-01-08T02:23:48,346 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742406_1582 (size=322274) 2025-01-08T02:23:48,346 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742406_1582 (size=322274) 2025-01-08T02:23:48,346 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742406_1582 (size=322274) 2025-01-08T02:23:48,356 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T02:23:48,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T02:23:48,357 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742407_1583 (size=1832290) 2025-01-08T02:23:48,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742408_1584 (size=30081) 2025-01-08T02:23:48,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742408_1584 (size=30081) 2025-01-08T02:23:48,363 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742408_1584 (size=30081) 2025-01-08T02:23:48,369 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742409_1585 (size=53616) 2025-01-08T02:23:48,369 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742409_1585 (size=53616) 2025-01-08T02:23:48,370 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742409_1585 (size=53616) 2025-01-08T02:23:48,375 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742410_1586 (size=29229) 2025-01-08T02:23:48,376 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742410_1586 (size=29229) 2025-01-08T02:23:48,376 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742410_1586 (size=29229) 2025-01-08T02:23:48,382 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742411_1587 (size=169089) 2025-01-08T02:23:48,382 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742411_1587 (size=169089) 2025-01-08T02:23:48,382 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742411_1587 (size=169089) 2025-01-08T02:23:48,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742412_1588 (size=451756) 2025-01-08T02:23:48,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742412_1588 (size=451756) 2025-01-08T02:23:48,390 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742412_1588 (size=451756) 2025-01-08T02:23:48,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742413_1589 (size=5175431) 2025-01-08T02:23:48,408 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742413_1589 (size=5175431) 2025-01-08T02:23:48,409 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742413_1589 (size=5175431) 2025-01-08T02:23:48,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742414_1590 (size=136454) 2025-01-08T02:23:48,420 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742414_1590 (size=136454) 2025-01-08T02:23:48,421 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742414_1590 (size=136454) 2025-01-08T02:23:48,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742415_1591 (size=907469) 2025-01-08T02:23:48,429 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742415_1591 (size=907469) 2025-01-08T02:23:48,430 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742415_1591 (size=907469) 2025-01-08T02:23:48,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(131): Registering adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:23:48,431 INFO [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(135): Registering RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp Metrics about Tables on a single HBase RegionServer 2025-01-08T02:23:48,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportWithChecksum 2025-01-08T02:23:48,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742416_1592 (size=3317408) 2025-01-08T02:23:48,444 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742416_1592 (size=3317408) 2025-01-08T02:23:48,445 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742416_1592 (size=3317408) 2025-01-08T02:23:48,452 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742417_1593 (size=503880) 2025-01-08T02:23:48,452 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742417_1593 (size=503880) 2025-01-08T02:23:48,453 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742417_1593 (size=503880) 2025-01-08T02:23:48,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T02:23:48,473 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T02:23:48,474 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742418_1594 (size=4695811) 2025-01-08T02:23:48,474 WARN [Time-limited test {}] mapreduce.JobResourceUploader(481): No job jar file set. User classes may not be found. See Job or Job#setJar(String). 2025-01-08T02:23:48,476 INFO [Time-limited test {}] snapshot.ExportSnapshot(658): Loading Snapshot 'snaptb0-testExportFileSystemStateWithSkipTmp' hfile list 2025-01-08T02:23:48,478 DEBUG [Time-limited test {}] snapshot.ExportSnapshot(751): export split=0 size=33.9 K 2025-01-08T02:23:48,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742419_1595 (size=770) 2025-01-08T02:23:48,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742419_1595 (size=770) 2025-01-08T02:23:48,484 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742419_1595 (size=770) 2025-01-08T02:23:48,489 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742420_1596 (size=15) 2025-01-08T02:23:48,490 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742420_1596 (size=15) 2025-01-08T02:23:48,490 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742420_1596 (size=15) 2025-01-08T02:23:48,501 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742421_1597 (size=305049) 2025-01-08T02:23:48,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742421_1597 (size=305049) 2025-01-08T02:23:48,502 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742421_1597 (size=305049) 2025-01-08T02:23:49,289 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(880): maximum-am-resource-percent is insufficient to start a single application in queue, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:23:49,289 WARN [SchedulerEventDispatcher:Event Processor {}] capacity.AbstractLeafQueue(913): maximum-am-resource-percent is insufficient to start a single application in queue for user, it is likely set too low. skipping enforcement to allow at least one application to start 2025-01-08T02:23:49,291 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0009_000001 (auth:SIMPLE) from 127.0.0.1:34842 2025-01-08T02:23:49,303 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0009/container_1736302816589_0009_01_000001/launch_container.sh] 2025-01-08T02:23:49,303 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0009/container_1736302816589_0009_01_000001/container_tokens] 2025-01-08T02:23:49,303 WARN [ContainersLauncher #0 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-1_1/usercache/jenkins/appcache/application_1736302816589_0009/container_1736302816589_0009_01_000001/sysfs] 2025-01-08T02:23:50,039 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0010_000001 (auth:SIMPLE) from 127.0.0.1:54548 2025-01-08T02:23:50,474 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:23:54,834 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0010_000001 (auth:SIMPLE) from 127.0.0.1:36376 2025-01-08T02:23:55,054 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742422_1598 (size=350747) 2025-01-08T02:23:55,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742422_1598 (size=350747) 2025-01-08T02:23:55,055 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742422_1598 (size=350747) 2025-01-08T02:23:57,068 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0010_000001 (auth:SIMPLE) from 127.0.0.1:59010 2025-01-08T02:23:59,158 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 2f4301e371bc30413722a847ac142fb1, had cached 0 bytes from a total of 6086 2025-01-08T02:23:59,158 DEBUG [HBase-Metrics2-1 {}] regionserver.MetricsRegionWrapperImpl$HRegionMetricsWrapperRunnable(324): Region 4c7db76a9413d28afb5a4d977c0db552, had cached 0 bytes from a total of 14463 2025-01-08T02:23:59,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742423_1599 (size=15309) 2025-01-08T02:23:59,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742423_1599 (size=15309) 2025-01-08T02:23:59,829 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742423_1599 (size=15309) 2025-01-08T02:23:59,879 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742424_1600 (size=8101) 2025-01-08T02:23:59,880 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742424_1600 (size=8101) 2025-01-08T02:23:59,880 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742424_1600 (size=8101) 2025-01-08T02:23:59,926 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742425_1601 (size=6176) 2025-01-08T02:23:59,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742425_1601 (size=6176) 2025-01-08T02:23:59,927 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742425_1601 (size=6176) 2025-01-08T02:23:59,966 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742426_1602 (size=5171) 2025-01-08T02:23:59,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742426_1602 (size=5171) 2025-01-08T02:23:59,967 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742426_1602 (size=5171) 2025-01-08T02:24:00,026 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742427_1603 (size=17473) 2025-01-08T02:24:00,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742427_1603 (size=17473) 2025-01-08T02:24:00,027 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742427_1603 (size=17473) 2025-01-08T02:24:00,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742428_1604 (size=476) 2025-01-08T02:24:00,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742428_1604 (size=476) 2025-01-08T02:24:00,041 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742428_1604 (size=476) 2025-01-08T02:24:00,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742429_1605 (size=17473) 2025-01-08T02:24:00,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742429_1605 (size=17473) 2025-01-08T02:24:00,079 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742429_1605 (size=17473) 2025-01-08T02:24:00,080 WARN [ContainersLauncher #3 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_1/usercache/jenkins/appcache/application_1736302816589_0010/container_1736302816589_0010_01_000002/launch_container.sh] 2025-01-08T02:24:00,081 WARN [ContainersLauncher #3 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_1/usercache/jenkins/appcache/application_1736302816589_0010/container_1736302816589_0010_01_000002/container_tokens] 2025-01-08T02:24:00,081 WARN [ContainersLauncher #3 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_1/usercache/jenkins/appcache/application_1736302816589_0010/container_1736302816589_0010_01_000002/sysfs] 2025-01-08T02:24:00,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742430_1606 (size=350747) 2025-01-08T02:24:00,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742430_1606 (size=350747) 2025-01-08T02:24:00,104 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742430_1606 (size=350747) 2025-01-08T02:24:00,117 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0010_000001 (auth:SIMPLE) from 127.0.0.1:59014 2025-01-08T02:24:01,627 INFO [Time-limited test {}] snapshot.ExportSnapshot(1207): Finalize the Snapshot Export 2025-01-08T02:24:01,627 INFO [Time-limited test {}] snapshot.ExportSnapshot(1218): Verify the exported snapshot's expiration status and integrity. 2025-01-08T02:24:01,636 INFO [Time-limited test {}] snapshot.ExportSnapshot(1224): Export Completed: snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,636 INFO [Time-limited test {}] snapshot.TestExportSnapshot(362): Exported snapshot 2025-01-08T02:24:01,636 INFO [Time-limited test {}] snapshot.TestExportSnapshot(373): Verified filesystem state 2025-01-08T02:24:01,636 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,637 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/.snapshotinfo 2025-01-08T02:24:01,637 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/data.manifest 2025-01-08T02:24:01,637 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(448): List files in DFS[DFSClient[clientName=DFSClient_NONMAPREDUCE_-1433654332_22, ugi=jenkins (auth:SIMPLE)]] in root hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp at hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,637 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/.snapshotinfo 2025-01-08T02:24:01,637 DEBUG [Time-limited test {}] snapshot.TestExportSnapshot(453): hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/export-test/export-1736303026542/.hbase-snapshot/snaptb0-testExportFileSystemStateWithSkipTmp/data.manifest 2025-01-08T02:24:01,643 INFO [Time-limited test {}] client.HBaseAdmin$18(967): Started disable of testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,644 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$13(2770): Client=jenkins//172.17.0.3 disable testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,644 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=209, state=RUNNABLE:DISABLE_TABLE_PREPARE; DisableTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,646 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T02:24:01,646 DEBUG [PEWorker-5 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736303041646"}]},"ts":"1736303041646"} 2025-01-08T02:24:01,647 INFO [PEWorker-5 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=DISABLING in hbase:meta 2025-01-08T02:24:01,649 INFO [PEWorker-5 {}] procedure.DisableTableProcedure(284): Set testtb-testExportFileSystemStateWithSkipTmp to state=DISABLING 2025-01-08T02:24:01,649 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=210, ppid=209, state=RUNNABLE:CLOSE_TABLE_REGIONS_SCHEDULE; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithSkipTmp}] 2025-01-08T02:24:01,650 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=211, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=947fe374ffc550ab7ab42fb17014d00e, UNASSIGN}, {pid=212, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=85b534583da52ed359fa76f762619b11, UNASSIGN}] 2025-01-08T02:24:01,651 INFO [PEWorker-1 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=211, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=947fe374ffc550ab7ab42fb17014d00e, UNASSIGN 2025-01-08T02:24:01,651 INFO [PEWorker-3 {}] procedure.MasterProcedureScheduler(786): Took xlock for pid=212, ppid=210, state=RUNNABLE:REGION_STATE_TRANSITION_CLOSE; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=85b534583da52ed359fa76f762619b11, UNASSIGN 2025-01-08T02:24:01,652 INFO [PEWorker-3 {}] assignment.RegionStateStore(202): pid=212 updating hbase:meta row=85b534583da52ed359fa76f762619b11, regionState=CLOSING, regionLocation=c4e5d52b9759,39911,1736302808913 2025-01-08T02:24:01,652 INFO [PEWorker-1 {}] assignment.RegionStateStore(202): pid=211 updating hbase:meta row=947fe374ffc550ab7ab42fb17014d00e, regionState=CLOSING, regionLocation=c4e5d52b9759,36743,1736302808837 2025-01-08T02:24:01,653 DEBUG [PEWorker-1 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:24:01,653 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=213, ppid=211, state=RUNNABLE; CloseRegionProcedure 947fe374ffc550ab7ab42fb17014d00e, server=c4e5d52b9759,36743,1736302808837}] 2025-01-08T02:24:01,653 DEBUG [PEWorker-3 {}] assignment.TransitRegionStateProcedure(338): Close region: isSplit: false: evictOnSplit: true: evictOnClose: false 2025-01-08T02:24:01,653 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1819): Initialized subprocedures=[{pid=214, ppid=212, state=RUNNABLE; CloseRegionProcedure 85b534583da52ed359fa76f762619b11, server=c4e5d52b9759,39911,1736302808913}] 2025-01-08T02:24:01,746 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T02:24:01,804 DEBUG [RSProcedureDispatcher-pool-0 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,39911,1736302808913 2025-01-08T02:24:01,804 DEBUG [RSProcedureDispatcher-pool-1 {}] master.ServerManager(801): New admin connection to c4e5d52b9759,36743,1736302808837 2025-01-08T02:24:01,804 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(124): Close 85b534583da52ed359fa76f762619b11 2025-01-08T02:24:01,804 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(124): Close 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(138): Unassign region: split region: false: evictCache: false 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1681): Closing 947fe374ffc550ab7ab42fb17014d00e, disabling compactions & flushes 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1681): Closing 85b534583da52ed359fa76f762619b11, disabling compactions & flushes 2025-01-08T02:24:01,805 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:24:01,805 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1703): Closing region testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1724): Waiting without time limit for close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. after waiting 0 ms 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1791): Acquired close lock on testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. after waiting 0 ms 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:24:01,805 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1801): Updates disabled for region testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:24:01,809 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:24:01,809 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:24:01,809 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:01,810 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:01,810 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e. 2025-01-08T02:24:01,810 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] regionserver.HRegion(1635): Region close journal for 947fe374ffc550ab7ab42fb17014d00e: 2025-01-08T02:24:01,810 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1922): Closed testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11. 2025-01-08T02:24:01,810 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] regionserver.HRegion(1635): Region close journal for 85b534583da52ed359fa76f762619b11: 2025-01-08T02:24:01,811 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=213}] handler.UnassignRegionHandler(170): Closed 947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:24:01,811 INFO [PEWorker-4 {}] assignment.RegionStateStore(202): pid=211 updating hbase:meta row=947fe374ffc550ab7ab42fb17014d00e, regionState=CLOSED 2025-01-08T02:24:01,811 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION, pid=214}] handler.UnassignRegionHandler(170): Closed 85b534583da52ed359fa76f762619b11 2025-01-08T02:24:01,812 INFO [PEWorker-2 {}] assignment.RegionStateStore(202): pid=212 updating hbase:meta row=85b534583da52ed359fa76f762619b11, regionState=CLOSED 2025-01-08T02:24:01,813 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=213, resume processing ppid=211 2025-01-08T02:24:01,813 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=213, ppid=211, state=SUCCESS; CloseRegionProcedure 947fe374ffc550ab7ab42fb17014d00e, server=c4e5d52b9759,36743,1736302808837 in 159 msec 2025-01-08T02:24:01,814 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=214, resume processing ppid=212 2025-01-08T02:24:01,814 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=214, ppid=212, state=SUCCESS; CloseRegionProcedure 85b534583da52ed359fa76f762619b11, server=c4e5d52b9759,39911,1736302808913 in 160 msec 2025-01-08T02:24:01,814 INFO [PEWorker-3 {}] procedure2.ProcedureExecutor(1480): Finished pid=211, ppid=210, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=947fe374ffc550ab7ab42fb17014d00e, UNASSIGN in 163 msec 2025-01-08T02:24:01,815 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=212, resume processing ppid=210 2025-01-08T02:24:01,815 INFO [PEWorker-1 {}] procedure2.ProcedureExecutor(1480): Finished pid=212, ppid=210, state=SUCCESS; TransitRegionStateProcedure table=testtb-testExportFileSystemStateWithSkipTmp, region=85b534583da52ed359fa76f762619b11, UNASSIGN in 164 msec 2025-01-08T02:24:01,816 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1976): Finished subprocedure pid=210, resume processing ppid=209 2025-01-08T02:24:01,816 INFO [PEWorker-5 {}] procedure2.ProcedureExecutor(1480): Finished pid=210, ppid=209, state=SUCCESS; CloseTableRegionsProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 166 msec 2025-01-08T02:24:01,817 DEBUG [PEWorker-4 {}] hbase.MetaTableAccessor(2113): Put {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":2,"tag":[],"timestamp":"1736303041817"}]},"ts":"1736303041817"} 2025-01-08T02:24:01,818 INFO [PEWorker-4 {}] hbase.MetaTableAccessor(1655): Updated tableName=testtb-testExportFileSystemStateWithSkipTmp, state=DISABLED in hbase:meta 2025-01-08T02:24:01,820 INFO [PEWorker-4 {}] procedure.DisableTableProcedure(296): Set testtb-testExportFileSystemStateWithSkipTmp to state=DISABLED 2025-01-08T02:24:01,821 INFO [PEWorker-4 {}] procedure2.ProcedureExecutor(1480): Finished pid=209, state=SUCCESS; DisableTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 176 msec 2025-01-08T02:24:01,947 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=209 2025-01-08T02:24:01,947 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DISABLE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 209 completed 2025-01-08T02:24:01,948 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.HMaster$5(2520): Client=jenkins//172.17.0.3 delete testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,949 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] procedure2.ProcedureExecutor(1098): Stored pid=215, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,949 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(103): Waiting for RIT for pid=215, state=RUNNABLE:DELETE_TABLE_PRE_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,949 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] access.PermissionStorage(259): Removing permissions of removed table testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,950 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(115): Deleting regions from filesystem for pid=215, state=RUNNABLE:DELETE_TABLE_CLEAR_FS_LAYOUT, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,951 INFO [RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=37031 {}] access.PermissionStorage(527): No permissions found in hbase:acl for acl entry testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,952 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:24:01,952 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11 2025-01-08T02:24:01,954 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,954 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,954 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,954 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/recovered.edits] 2025-01-08T02:24:01,954 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDataChanged, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,954 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/cf, FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/recovered.edits] 2025-01-08T02:24:01,954 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T02:24:01,955 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T02:24:01,955 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T02:24:01,955 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testtb-testExportFileSystemStateWithSkipTmp with data PBUF 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/acl/testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,956 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/acl 2025-01-08T02:24:01,957 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=215 2025-01-08T02:24:01,957 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:24:01,957 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:24:01,957 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:24:01,957 DEBUG [zk-permission-watcher-pool-0 {}] access.ZKPermissionWatcher(245): Updating permissions cache from testExportExpiredSnapshot with data PBUF\x0A?\x0A\x07jenkins\x124\x08\x03"0\x0A$\x0A\x07default\x12\x19testExportExpiredSnapshot \x00 \x01 \x02 \x03 \x04 2025-01-08T02:24:01,959 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/cf/2ba935ee465643b5a27d9ad9371f2d13 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/cf/2ba935ee465643b5a27d9ad9371f2d13 2025-01-08T02:24:01,959 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/cf/d4958f48a1a64f829999f946a2049531 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/cf/d4958f48a1a64f829999f946a2049531 2025-01-08T02:24:01,961 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11/recovered.edits/9.seqid 2025-01-08T02:24:01,961 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/recovered.edits/9.seqid to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e/recovered.edits/9.seqid 2025-01-08T02:24:01,962 DEBUG [HFileArchiver-22 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/85b534583da52ed359fa76f762619b11 2025-01-08T02:24:01,962 DEBUG [HFileArchiver-21 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testtb-testExportFileSystemStateWithSkipTmp/947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:24:01,962 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(313): Archived testtb-testExportFileSystemStateWithSkipTmp regions 2025-01-08T02:24:01,962 DEBUG [PEWorker-2 {}] backup.HFileArchiver(131): ARCHIVING hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a 2025-01-08T02:24:01,963 DEBUG [PEWorker-2 {}] backup.HFileArchiver(159): Archiving [FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf] 2025-01-08T02:24:01,966 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108da8c61020b7b45279d4a0198c20f55f4_85b534583da52ed359fa76f762619b11 to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/c4ca4238a0b923820dcc509a6f75849b20250108da8c61020b7b45279d4a0198c20f55f4_85b534583da52ed359fa76f762619b11 2025-01-08T02:24:01,967 DEBUG [PEWorker-2 {}] backup.HFileArchiver(596): Archived from FileablePath, hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e20250108d27cad870ca5438fae93a54470fb486f_947fe374ffc550ab7ab42fb17014d00e to hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/archive/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a/cf/d41d8cd98f00b204e9800998ecf8427e20250108d27cad870ca5438fae93a54470fb486f_947fe374ffc550ab7ab42fb17014d00e 2025-01-08T02:24:01,967 DEBUG [PEWorker-2 {}] backup.HFileArchiver(610): Deleted hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/mobdir/data/default/testtb-testExportFileSystemStateWithSkipTmp/e9516f04c7d2974b21addb190acc6c0a 2025-01-08T02:24:01,969 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(120): Deleting regions from META for pid=215, state=RUNNABLE:DELETE_TABLE_REMOVE_FROM_META, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,971 WARN [PEWorker-2 {}] procedure.DeleteTableProcedure(370): Deleting some vestigial 2 rows of testtb-testExportFileSystemStateWithSkipTmp from hbase:meta 2025-01-08T02:24:01,973 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(407): Removing 'testtb-testExportFileSystemStateWithSkipTmp' descriptor. 2025-01-08T02:24:01,973 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(126): Deleting assignment state for pid=215, state=RUNNABLE:DELETE_TABLE_UNASSIGN_REGIONS, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,973 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(397): Removing 'testtb-testExportFileSystemStateWithSkipTmp' from region states. 2025-01-08T02:24:01,974 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736303041973"}]},"ts":"9223372036854775807"} 2025-01-08T02:24:01,974 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.","families":{"info":[{"qualifier":"","vlen":0,"tag":[],"timestamp":"1736303041973"}]},"ts":"9223372036854775807"} 2025-01-08T02:24:01,975 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1808): Deleted 2 regions from META 2025-01-08T02:24:01,975 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(1809): Deleted regions: [{ENCODED => 947fe374ffc550ab7ab42fb17014d00e, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,,1736303025252.947fe374ffc550ab7ab42fb17014d00e.', STARTKEY => '', ENDKEY => '1'}, {ENCODED => 85b534583da52ed359fa76f762619b11, NAME => 'testtb-testExportFileSystemStateWithSkipTmp,1,1736303025252.85b534583da52ed359fa76f762619b11.', STARTKEY => '1', ENDKEY => ''}] 2025-01-08T02:24:01,975 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(401): Marking 'testtb-testExportFileSystemStateWithSkipTmp' as deleted. 2025-01-08T02:24:01,975 DEBUG [PEWorker-2 {}] hbase.MetaTableAccessor(2113): Delete {"totalColumns":1,"row":"testtb-testExportFileSystemStateWithSkipTmp","families":{"table":[{"qualifier":"state","vlen":0,"tag":[],"timestamp":"1736303041975"}]},"ts":"9223372036854775807"} 2025-01-08T02:24:01,976 INFO [PEWorker-2 {}] hbase.MetaTableAccessor(1678): Deleted table testtb-testExportFileSystemStateWithSkipTmp state from META 2025-01-08T02:24:01,978 DEBUG [PEWorker-2 {}] procedure.DeleteTableProcedure(133): Finished pid=215, state=RUNNABLE:DELETE_TABLE_POST_OPERATION, locked=true; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:01,979 INFO [PEWorker-2 {}] procedure2.ProcedureExecutor(1480): Finished pid=215, state=SUCCESS; DeleteTableProcedure table=testtb-testExportFileSystemStateWithSkipTmp in 30 msec 2025-01-08T02:24:02,057 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(1305): Checking to see if procedure is done pid=215 2025-01-08T02:24:02,057 INFO [Time-limited test {}] client.HBaseAdmin$TableFuture(3751): Operation: DELETE, Table Name: default:testtb-testExportFileSystemStateWithSkipTmp, procId: 215 completed 2025-01-08T02:24:02,062 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "emptySnaptb0-testExportFileSystemStateWithSkipTmp" 2025-01-08T02:24:02,063 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: emptySnaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:02,065 INFO [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] master.MasterRpcServices(764): Client=jenkins//172.17.0.3 delete name: "snaptb0-testExportFileSystemStateWithSkipTmp" 2025-01-08T02:24:02,066 DEBUG [RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063 {}] snapshot.SnapshotManager(380): Deleting snapshot: snaptb0-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:02,085 INFO [Time-limited test {}] hbase.ResourceChecker(175): after: snapshot.TestMobSecureExportSnapshot#testExportFileSystemStateWithSkipTmp Thread=807 (was 802) Potentially hanging thread: IPC Client (831077991) connection to localhost/127.0.0.1:42267 from jenkins java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:36580 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-52 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:60880 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #18 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: Thread-7785 java.base@17.0.11/java.io.FileInputStream.readBytes(Native Method) java.base@17.0.11/java.io.FileInputStream.read(FileInputStream.java:276) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:282) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) java.base@17.0.11/sun.nio.cs.StreamDecoder.readBytes(StreamDecoder.java:281) java.base@17.0.11/sun.nio.cs.StreamDecoder.implRead(StreamDecoder.java:324) java.base@17.0.11/sun.nio.cs.StreamDecoder.read(StreamDecoder.java:189) java.base@17.0.11/java.io.InputStreamReader.read(InputStreamReader.java:177) java.base@17.0.11/java.io.BufferedReader.fill(BufferedReader.java:162) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:329) java.base@17.0.11/java.io.BufferedReader.readLine(BufferedReader.java:396) app//org.apache.hadoop.util.Shell$1.run(Shell.java:1025) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-50 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-49 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: process reaper (pid 17374) java.base@17.0.11/java.lang.ProcessHandleImpl.waitForProcessExit0(Native Method) java.base@17.0.11/java.lang.ProcessHandleImpl$1.run(ProcessHandleImpl.java:150) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: hconnection-0x35e289a5-shared-pool-51 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: IPC Parameter Sending Thread for localhost/127.0.0.1:42267 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-742936593_1 at /127.0.0.1:36574 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: ApplicationMasterLauncher #17 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-22 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-1433654332_22 at /127.0.0.1:33476 [Waiting for operation #5] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: HFileArchiver-21 java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Potentially hanging thread: DataXceiver for client DFSClient_NONMAPREDUCE_-742936593_1 at /127.0.0.1:60856 [Waiting for operation #2] java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select(SocketIOWithTimeout.java:335) app//org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:156) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) app//org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:263) java.base@17.0.11/java.io.DataInputStream.readUnsignedShort(DataInputStream.java:334) java.base@17.0.11/java.io.DataInputStream.readShort(DataInputStream.java:312) app//org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.readOp(Receiver.java:72) app//org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:273) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) - Thread LEAK? -, OpenFileDescriptor=809 (was 813), MaxFileDescriptor=1048576 (was 1048576), SystemLoadAverage=399 (was 464), ProcessCount=18 (was 18), AvailableMemoryMB=669 (was 784) 2025-01-08T02:24:02,085 WARN [Time-limited test {}] hbase.ResourceChecker(130): Thread=807 is superior to 500 2025-01-08T02:24:02,086 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2861): Stopping mini mapreduce cluster... 2025-01-08T02:24:02,092 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@3c91a94d{node,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T02:24:02,095 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@402c953e{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:24:02,095 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:24:02,095 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@66f21560{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T02:24:02,095 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@a5b7a6f{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:24:06,185 INFO [Socket Reader #1 for port 0 {}] ipc.Server$Connection(2290): Auth successful for appattempt_1736302816589_0010_000001 (auth:SIMPLE) from 127.0.0.1:37236 2025-01-08T02:24:06,196 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_1/usercache/jenkins/appcache/application_1736302816589_0010/container_1736302816589_0010_01_000001/launch_container.sh] 2025-01-08T02:24:06,196 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_1/usercache/jenkins/appcache/application_1736302816589_0010/container_1736302816589_0010_01_000001/container_tokens] 2025-01-08T02:24:06,196 WARN [ContainersLauncher #2 {}] nodemanager.DefaultContainerExecutor(697): delete returned false for path: [/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/MiniMRCluster_812930379/yarn-5760119216/MiniMRCluster_812930379-localDir-nm-0_1/usercache/jenkins/appcache/application_1736302816589_0010/container_1736302816589_0010_01_000001/sysfs] 2025-01-08T02:24:06,390 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:24:07,330 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:24:08,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testtb-testExportFileSystemStateWithSkipTmp 2025-01-08T02:24:13,933 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:24:19,107 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@4cc33919{node,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/node} 2025-01-08T02:24:19,108 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@67f55a9f{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:24:19,108 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:24:19,108 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6c4f46cc{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T02:24:19,108 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6346f6c{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:24:36,116 ERROR [Thread[Thread-416,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T02:24:36,116 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@634e8a70{cluster,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/cluster} 2025-01-08T02:24:36,117 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1dac9f1d{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:24:36,117 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:24:36,117 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@17d5de37{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T02:24:36,117 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@520a3669{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:24:36,121 WARN [ApplicationMaster Launcher {}] amlauncher.ApplicationMasterLauncher$LauncherThread(122): org.apache.hadoop.yarn.server.resourcemanager.amlauncher.ApplicationMasterLauncher$LauncherThread interrupted. Returning. 2025-01-08T02:24:36,127 ERROR [SchedulerEventDispatcher:Event Processor {}] event.EventDispatcher$EventProcessor(72): Returning, interrupted : java.lang.InterruptedException 2025-01-08T02:24:36,127 ERROR [ResourceManager Event Processor Monitor {}] resourcemanager.ResourceManager$SchedulerEventDispatcher$EventProcessorMonitor(1193): Returning, interrupted : java.lang.InterruptedException: sleep interrupted 2025-01-08T02:24:36,129 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741830_1006 (size=952517) 2025-01-08T02:24:36,129 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741830_1006 (size=952517) 2025-01-08T02:24:36,129 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741830_1006 (size=952517) 2025-01-08T02:24:36,134 ERROR [Thread[Thread-439,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T02:24:36,137 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@64ae69bd{jobhistory,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/jobhistory} 2025-01-08T02:24:36,138 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@1fac6a92{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:24:36,138 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:24:36,138 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@79f1c1ff{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-yarn-common/3.4.1/hadoop-yarn-common-3.4.1.jar!/webapps/static,STOPPED} 2025-01-08T02:24:36,138 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@4d920d82{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:24:36,139 ERROR [Thread[Thread-398,5,FailOnTimeoutGroup] {}] delegation.AbstractDelegationTokenSecretManager$ExpiredTokenRemover(852): ExpiredTokenRemover received java.lang.InterruptedException: sleep interrupted 2025-01-08T02:24:36,140 INFO [Time-limited test {}] hbase.HBaseTestingUtility(2864): Mini mapreduce cluster stopped 2025-01-08T02:24:36,140 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1340): Shutting down minicluster 2025-01-08T02:24:36,140 INFO [Time-limited test {}] client.ConnectionImplementation(2127): Closing master protocol: MasterService 2025-01-08T02:24:36,140 DEBUG [Time-limited test {}] zookeeper.ReadOnlyZKClient(407): Close zookeeper connection 0x5dda4b49 to 127.0.0.1:61593 2025-01-08T02:24:36,140 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,140 DEBUG [Time-limited test {}] util.JVMClusterUtil(237): Shutting down HBase Cluster 2025-01-08T02:24:36,140 DEBUG [Time-limited test {}] util.JVMClusterUtil(257): Found active master hash=2129446819, stopped=false 2025-01-08T02:24:36,140 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,141 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T02:24:36,141 INFO [Time-limited test {}] master.ServerManager(987): Cluster shutdown requested of master=c4e5d52b9759,41063,1736302807767 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/running 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:24:36,142 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:24:36,143 INFO [Time-limited test {}] procedure2.ProcedureExecutor(700): Stopping 2025-01-08T02:24:36,143 DEBUG [Time-limited test {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,143 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:24:36,143 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:24:36,143 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'c4e5d52b9759,37031,1736302808764' ***** 2025-01-08T02:24:36,143 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:24:36,143 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,143 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T02:24:36,143 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'c4e5d52b9759,36743,1736302808837' ***** 2025-01-08T02:24:36,143 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,143 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T02:24:36,143 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T02:24:36,143 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/running 2025-01-08T02:24:36,143 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T02:24:36,144 INFO [RS:1;c4e5d52b9759:36743 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T02:24:36,144 INFO [RS:0;c4e5d52b9759:37031 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T02:24:36,144 INFO [RS:0;c4e5d52b9759:37031 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T02:24:36,144 INFO [RS:1;c4e5d52b9759:36743 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T02:24:36,144 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T02:24:36,144 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T02:24:36,144 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(3579): Received CLOSE for 2f4301e371bc30413722a847ac142fb1 2025-01-08T02:24:36,144 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(3579): Received CLOSE for 4984d49fea614490cee04f171891b535 2025-01-08T02:24:36,144 INFO [Time-limited test {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'c4e5d52b9759,39911,1736302808913' ***** 2025-01-08T02:24:36,144 DEBUG [Time-limited test {}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,144 INFO [Time-limited test {}] regionserver.HRegionServer(2575): STOPPED: Shutdown requested 2025-01-08T02:24:36,144 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1224): stopping server c4e5d52b9759,36743,1736302808837 2025-01-08T02:24:36,144 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HeapMemoryManager(220): Stopping 2025-01-08T02:24:36,144 DEBUG [RS:1;c4e5d52b9759:36743 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,144 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(3579): Received CLOSE for 8170840180427f0da936e0809aa0d2c2 2025-01-08T02:24:36,144 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1224): stopping server c4e5d52b9759,37031,1736302808764 2025-01-08T02:24:36,144 INFO [RS:2;c4e5d52b9759:39911 {}] flush.RegionServerFlushTableProcedureManager(119): Stopping region server flush procedure manager gracefully. 2025-01-08T02:24:36,144 INFO [RS:2;c4e5d52b9759:39911 {}] snapshot.RegionServerSnapshotManager(137): Stopping RegionServerSnapshotManager gracefully. 2025-01-08T02:24:36,144 DEBUG [RS:0;c4e5d52b9759:37031 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,144 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T02:24:36,144 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T02:24:36,144 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T02:24:36,144 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1599): Waiting on 2 regions to close 2025-01-08T02:24:36,145 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(3579): Received CLOSE for 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:24:36,145 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1603): Online Regions={2f4301e371bc30413722a847ac142fb1=testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1., 8170840180427f0da936e0809aa0d2c2=hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2.} 2025-01-08T02:24:36,145 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1224): stopping server c4e5d52b9759,39911,1736302808913 2025-01-08T02:24:36,145 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(3579): Received CLOSE for 1588230740 2025-01-08T02:24:36,145 DEBUG [RS:2;c4e5d52b9759:39911 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,145 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1599): Waiting on 1 regions to close 2025-01-08T02:24:36,145 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1603): Online Regions={4c7db76a9413d28afb5a4d977c0db552=testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552.} 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 2f4301e371bc30413722a847ac142fb1, disabling compactions & flushes 2025-01-08T02:24:36,145 INFO [MemStoreFlusher.0 {}] regionserver.MemStoreFlusher$FlushHandler(373): MemStoreFlusher.0 exiting 2025-01-08T02:24:36,145 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. after waiting 0 ms 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:24:36,145 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1599): Waiting on 2 regions to close 2025-01-08T02:24:36,145 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1603): Online Regions={4984d49fea614490cee04f171891b535=hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535., 1588230740=hbase:meta,,1.1588230740} 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1681): Closing 1588230740, disabling compactions & flushes 2025-01-08T02:24:36,145 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1703): Closing region hbase:meta,,1.1588230740 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:meta,,1.1588230740 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1791): Acquired close lock on hbase:meta,,1.1588230740 after waiting 0 ms 2025-01-08T02:24:36,145 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1801): Updates disabled for region hbase:meta,,1.1588230740 2025-01-08T02:24:36,146 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(2837): Flushing 1588230740 3/3 column families, dataSize=68.66 KB heapSize=109 KB 2025-01-08T02:24:36,147 INFO [regionserver/c4e5d52b9759:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 4984d49fea614490cee04f171891b535, disabling compactions & flushes 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 4c7db76a9413d28afb5a4d977c0db552, disabling compactions & flushes 2025-01-08T02:24:36,148 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:24:36,148 DEBUG [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1629): Waiting on 4c7db76a9413d28afb5a4d977c0db552 2025-01-08T02:24:36,148 DEBUG [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1629): Waiting on 1588230740, 4984d49fea614490cee04f171891b535 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. after waiting 0 ms 2025-01-08T02:24:36,148 DEBUG [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1629): Waiting on 2f4301e371bc30413722a847ac142fb1, 8170840180427f0da936e0809aa0d2c2 2025-01-08T02:24:36,148 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. after waiting 0 ms 2025-01-08T02:24:36,148 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:24:36,148 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 4984d49fea614490cee04f171891b535 1/1 column families, dataSize=78 B heapSize=488 B 2025-01-08T02:24:36,153 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/2f4301e371bc30413722a847ac142fb1/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/default/testExportExpiredSnapshot/4c7db76a9413d28afb5a4d977c0db552/recovered.edits/8.seqid, newMaxSeqId=8, maxSeqId=1 2025-01-08T02:24:36,154 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 2f4301e371bc30413722a847ac142fb1: 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed testExportExpiredSnapshot,,1736302948799.2f4301e371bc30413722a847ac142fb1. 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1681): Closing 8170840180427f0da936e0809aa0d2c2, disabling compactions & flushes 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,154 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1703): Closing region hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:24:36,154 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1724): Waiting without time limit for close lock on hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1791): Acquired close lock on hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. after waiting 0 ms 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 4c7db76a9413d28afb5a4d977c0db552: 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1801): Updates disabled for region hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:24:36,154 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552. 2025-01-08T02:24:36,154 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(2837): Flushing 8170840180427f0da936e0809aa0d2c2 1/1 column families, dataSize=1.38 KB heapSize=3.33 KB 2025-01-08T02:24:36,155 INFO [regionserver/c4e5d52b9759:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T02:24:36,166 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535/.tmp/info/1018def9651f483a9e68b0659e81ceae is 45, key is default/info:d/1736302812475/Put/seqid=0 2025-01-08T02:24:36,168 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2/.tmp/l/a6adae570d2d44b5820c4f742584da1d is 74, key is testtb-testExportFileSystemStateWithMergeRegion-1/l:/1736302946113/DeleteFamily/seqid=0 2025-01-08T02:24:36,170 INFO [regionserver/c4e5d52b9759:0.leaseChecker {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T02:24:36,174 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/info/b5387ab8a88848bdbee6ed9cbd035ae8 is 173, key is testExportExpiredSnapshot,1,1736302948799.4c7db76a9413d28afb5a4d977c0db552./info:regioninfo/1736302949174/Put/seqid=0 2025-01-08T02:24:36,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742432_1608 (size=5695) 2025-01-08T02:24:36,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742432_1608 (size=5695) 2025-01-08T02:24:36,174 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742432_1608 (size=5695) 2025-01-08T02:24:36,175 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=1.38 KB at sequenceid=27 (bloomFilter=false), to=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2/.tmp/l/a6adae570d2d44b5820c4f742584da1d 2025-01-08T02:24:36,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742431_1607 (size=5037) 2025-01-08T02:24:36,176 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742431_1607 (size=5037) 2025-01-08T02:24:36,177 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742431_1607 (size=5037) 2025-01-08T02:24:36,177 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=78 B at sequenceid=6 (bloomFilter=true), to=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535/.tmp/info/1018def9651f483a9e68b0659e81ceae 2025-01-08T02:24:36,180 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742433_1609 (size=15630) 2025-01-08T02:24:36,180 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for a6adae570d2d44b5820c4f742584da1d 2025-01-08T02:24:36,180 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742433_1609 (size=15630) 2025-01-08T02:24:36,181 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742433_1609 (size=15630) 2025-01-08T02:24:36,181 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2/.tmp/l/a6adae570d2d44b5820c4f742584da1d as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2/l/a6adae570d2d44b5820c4f742584da1d 2025-01-08T02:24:36,181 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=60.26 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/info/b5387ab8a88848bdbee6ed9cbd035ae8 2025-01-08T02:24:36,183 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535/.tmp/info/1018def9651f483a9e68b0659e81ceae as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535/info/1018def9651f483a9e68b0659e81ceae 2025-01-08T02:24:36,185 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.StoreFileReader(539): Loaded Delete Family Bloom (CompoundBloomFilter) metadata for a6adae570d2d44b5820c4f742584da1d 2025-01-08T02:24:36,185 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2/l/a6adae570d2d44b5820c4f742584da1d, entries=12, sequenceid=27, filesize=5.6 K 2025-01-08T02:24:36,186 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~1.38 KB/1412, heapSize ~3.31 KB/3392, currentSize=0 B/0 for 8170840180427f0da936e0809aa0d2c2 in 32ms, sequenceid=27, compaction requested=false 2025-01-08T02:24:36,187 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535/info/1018def9651f483a9e68b0659e81ceae, entries=2, sequenceid=6, filesize=4.9 K 2025-01-08T02:24:36,188 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(3040): Finished flush of dataSize ~78 B/78, heapSize ~472 B/472, currentSize=0 B/0 for 4984d49fea614490cee04f171891b535 in 40ms, sequenceid=6, compaction requested=false 2025-01-08T02:24:36,189 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/acl/8170840180427f0da936e0809aa0d2c2/recovered.edits/30.seqid, newMaxSeqId=30, maxSeqId=1 2025-01-08T02:24:36,190 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,190 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:24:36,190 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 8170840180427f0da936e0809aa0d2c2: 2025-01-08T02:24:36,190 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:acl,,1736302812551.8170840180427f0da936e0809aa0d2c2. 2025-01-08T02:24:36,191 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/namespace/4984d49fea614490cee04f171891b535/recovered.edits/9.seqid, newMaxSeqId=9, maxSeqId=1 2025-01-08T02:24:36,191 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,191 INFO [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1922): Closed hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:24:36,191 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] regionserver.HRegion(1635): Region close journal for 4984d49fea614490cee04f171891b535: 2025-01-08T02:24:36,191 DEBUG [RS_CLOSE_REGION-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_REGION}] handler.CloseRegionHandler(117): Closed hbase:namespace,,1736302811497.4984d49fea614490cee04f171891b535. 2025-01-08T02:24:36,200 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/rep_barrier/b6e03bd0df344d1b968fe44b3ecd2a22 is 133, key is testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598./rep_barrier:/1736302946131/DeleteFamily/seqid=0 2025-01-08T02:24:36,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742434_1610 (size=8007) 2025-01-08T02:24:36,204 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742434_1610 (size=8007) 2025-01-08T02:24:36,205 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742434_1610 (size=8007) 2025-01-08T02:24:36,205 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=2.34 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/rep_barrier/b6e03bd0df344d1b968fe44b3ecd2a22 2025-01-08T02:24:36,223 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] hfile.HFileWriterImpl(821): Len of the biggest cell in hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/table/0debc3ce0b8941be8b962cedc19cf408 is 127, key is testtb-testExportFileSystemStateWithMergeRegion-1,,1736302927624.a25e31c91636f9ea9e4f60f60a12d598./table:/1736302946131/DeleteFamily/seqid=0 2025-01-08T02:24:36,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073742435_1611 (size=8861) 2025-01-08T02:24:36,227 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073742435_1611 (size=8861) 2025-01-08T02:24:36,228 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073742435_1611 (size=8861) 2025-01-08T02:24:36,228 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.DefaultStoreFlusher(81): Flushed memstore data size=6.06 KB at sequenceid=202 (bloomFilter=true), to=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/table/0debc3ce0b8941be8b962cedc19cf408 2025-01-08T02:24:36,232 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/info/b5387ab8a88848bdbee6ed9cbd035ae8 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/info/b5387ab8a88848bdbee6ed9cbd035ae8 2025-01-08T02:24:36,236 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/info/b5387ab8a88848bdbee6ed9cbd035ae8, entries=84, sequenceid=202, filesize=15.3 K 2025-01-08T02:24:36,237 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/rep_barrier/b6e03bd0df344d1b968fe44b3ecd2a22 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/rep_barrier/b6e03bd0df344d1b968fe44b3ecd2a22 2025-01-08T02:24:36,240 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/rep_barrier/b6e03bd0df344d1b968fe44b3ecd2a22, entries=21, sequenceid=202, filesize=7.8 K 2025-01-08T02:24:36,241 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegionFileSystem(442): Committing hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/.tmp/table/0debc3ce0b8941be8b962cedc19cf408 as hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/table/0debc3ce0b8941be8b962cedc19cf408 2025-01-08T02:24:36,244 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HStore$StoreFlusherImpl(1989): Added hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/table/0debc3ce0b8941be8b962cedc19cf408, entries=38, sequenceid=202, filesize=8.7 K 2025-01-08T02:24:36,244 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(3040): Finished flush of dataSize ~68.66 KB/70312, heapSize ~108.95 KB/111568, currentSize=0 B/0 for 1588230740 in 99ms, sequenceid=202, compaction requested=false 2025-01-08T02:24:36,247 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] wal.WALSplitUtil(409): Wrote file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/data/hbase/meta/1588230740/recovered.edits/205.seqid, newMaxSeqId=205, maxSeqId=1 2025-01-08T02:24:36,248 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:36,248 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] coprocessor.CoprocessorHost(310): Stop coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T02:24:36,248 INFO [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1922): Closed hbase:meta,,1.1588230740 2025-01-08T02:24:36,248 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] regionserver.HRegion(1635): Region close journal for 1588230740: 2025-01-08T02:24:36,248 DEBUG [RS_CLOSE_META-regionserver/c4e5d52b9759:0-0 {event_type=M_RS_CLOSE_META}] handler.CloseRegionHandler(117): Closed hbase:meta,,1.1588230740 2025-01-08T02:24:36,348 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1250): stopping server c4e5d52b9759,36743,1736302808837; all regions closed. 2025-01-08T02:24:36,348 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1250): stopping server c4e5d52b9759,39911,1736302808913; all regions closed. 2025-01-08T02:24:36,348 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1250): stopping server c4e5d52b9759,37031,1736302808764; all regions closed. 2025-01-08T02:24:36,352 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741834_1010 (size=14710) 2025-01-08T02:24:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741834_1010 (size=14710) 2025-01-08T02:24:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741834_1010 (size=14710) 2025-01-08T02:24:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741836_1012 (size=80694) 2025-01-08T02:24:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741835_1011 (size=13131) 2025-01-08T02:24:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741835_1011 (size=13131) 2025-01-08T02:24:36,353 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741836_1012 (size=80694) 2025-01-08T02:24:36,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741836_1012 (size=80694) 2025-01-08T02:24:36,354 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741835_1011 (size=13131) 2025-01-08T02:24:36,356 DEBUG [RS:2;c4e5d52b9759:39911 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs 2025-01-08T02:24:36,356 INFO [RS:2;c4e5d52b9759:39911 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL c4e5d52b9759%2C39911%2C1736302808913:(num 1736302810855) 2025-01-08T02:24:36,356 DEBUG [RS:2;c4e5d52b9759:39911 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,356 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T02:24:36,356 DEBUG [RS:0;c4e5d52b9759:37031 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs 2025-01-08T02:24:36,356 INFO [RS:0;c4e5d52b9759:37031 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL c4e5d52b9759%2C37031%2C1736302808764:(num 1736302810855) 2025-01-08T02:24:36,356 DEBUG [RS:0;c4e5d52b9759:37031 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,356 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T02:24:36,356 DEBUG [RS:1;c4e5d52b9759:36743 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs 2025-01-08T02:24:36,356 INFO [RS:1;c4e5d52b9759:36743 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL c4e5d52b9759%2C36743%2C1736302808837.meta:.meta(num 1736302811254) 2025-01-08T02:24:36,356 INFO [RS:2;c4e5d52b9759:39911 {}] hbase.ChoreService(370): Chore service for: regionserver/c4e5d52b9759:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2025-01-08T02:24:36,356 INFO [RS:0;c4e5d52b9759:37031 {}] hbase.ChoreService(370): Chore service for: regionserver/c4e5d52b9759:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2025-01-08T02:24:36,357 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T02:24:36,357 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.CompactSplit(469): Waiting for Split Thread to finish... 2025-01-08T02:24:36,357 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T02:24:36,357 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.CompactSplit(469): Waiting for Large Compaction Thread to finish... 2025-01-08T02:24:36,357 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T02:24:36,357 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.CompactSplit(469): Waiting for Small Compaction Thread to finish... 2025-01-08T02:24:36,357 INFO [regionserver/c4e5d52b9759:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T02:24:36,357 INFO [regionserver/c4e5d52b9759:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T02:24:36,357 INFO [RS:0;c4e5d52b9759:37031 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:37031 2025-01-08T02:24:36,357 WARN [BootstrapNodeManager {}] regionserver.BootstrapNodeManager(140): failed to get live region servers from master org.apache.hadoop.hbase.ipc.StoppedRpcClientException: Call to address=c4e5d52b9759:41063 failed on local exception: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) ~[?:?] at jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) ~[?:?] at java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) ~[?:?] at java.lang.reflect.Constructor.newInstance(Constructor.java:480) ~[?:?] at org.apache.hadoop.hbase.ipc.IPCUtil.wrapException(IPCUtil.java:237) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.onCallFinished(AbstractRpcClient.java:395) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$100(AbstractRpcClient.java:94) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:430) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$3.run(AbstractRpcClient.java:425) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.callComplete(Call.java:116) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.Call.setException(Call.java:131) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:451) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:336) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.access$200(AbstractRpcClient.java:94) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:606) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.getLiveRegionServers(RegionServerStatusProtos.java:17362) ~[hbase-protocol-shaded-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.client.ConnectionImplementation.getLiveRegionServers(ConnectionImplementation.java:2376) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.BootstrapNodeManager.getFromMaster(BootstrapNodeManager.java:138) ~[classes/:?] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?] at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[?:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.ipc.StoppedRpcClientException at org.apache.hadoop.hbase.ipc.AbstractRpcClient.getConnection(AbstractRpcClient.java:366) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callMethod(AbstractRpcClient.java:448) ~[hbase-client-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] ... 12 more 2025-01-08T02:24:36,358 INFO [RS:2;c4e5d52b9759:39911 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:39911 2025-01-08T02:24:36,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41199 is added to blk_1073741833_1009 (size=10433) 2025-01-08T02:24:36,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:41871 is added to blk_1073741833_1009 (size=10433) 2025-01-08T02:24:36,360 INFO [Block report processor {}] blockmanagement.BlockManager(3777): BLOCK* addStoredBlock: 127.0.0.1:43235 is added to blk_1073741833_1009 (size=10433) 2025-01-08T02:24:36,362 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/c4e5d52b9759,39911,1736302808913 2025-01-08T02:24:36,362 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T02:24:36,363 DEBUG [RS:1;c4e5d52b9759:36743 {}] wal.AbstractFSWAL(1071): Moved 1 WAL file(s) to /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/oldWALs 2025-01-08T02:24:36,363 INFO [RS:1;c4e5d52b9759:36743 {}] wal.AbstractFSWAL(1074): Closed WAL: AsyncFSWAL c4e5d52b9759%2C36743%2C1736302808837:(num 1736302810854) 2025-01-08T02:24:36,363 DEBUG [RS:1;c4e5d52b9759:36743 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,363 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.LeaseManager(133): Closed leases 2025-01-08T02:24:36,363 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/c4e5d52b9759,37031,1736302808764 2025-01-08T02:24:36,363 INFO [RS:1;c4e5d52b9759:36743 {}] hbase.ChoreService(370): Chore service for: regionserver/c4e5d52b9759:0 had [ScheduledChore name=CompactionThroughputTuner, period=60000, unit=MILLISECONDS, ScheduledChore name=BrokenStoreFileCleaner, period=21600000, unit=MILLISECONDS] on shutdown 2025-01-08T02:24:36,363 INFO [RS:1;c4e5d52b9759:36743 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:36743 2025-01-08T02:24:36,363 INFO [regionserver/c4e5d52b9759:0.logRoller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T02:24:36,363 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [c4e5d52b9759,37031,1736302808764] 2025-01-08T02:24:36,364 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing c4e5d52b9759,37031,1736302808764; numProcessing=1 2025-01-08T02:24:36,366 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/c4e5d52b9759,37031,1736302808764 already deleted, retry=false 2025-01-08T02:24:36,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase/rs 2025-01-08T02:24:36,366 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; c4e5d52b9759,37031,1736302808764 expired; onlineServers=2 2025-01-08T02:24:36,366 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/rs/c4e5d52b9759,36743,1736302808837 2025-01-08T02:24:36,366 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [c4e5d52b9759,39911,1736302808913] 2025-01-08T02:24:36,366 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing c4e5d52b9759,39911,1736302808913; numProcessing=2 2025-01-08T02:24:36,367 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/c4e5d52b9759,39911,1736302808913 already deleted, retry=false 2025-01-08T02:24:36,367 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; c4e5d52b9759,39911,1736302808913 expired; onlineServers=1 2025-01-08T02:24:36,368 INFO [RegionServerTracker-0 {}] master.RegionServerTracker(171): RegionServer ephemeral node deleted, processing expiration [c4e5d52b9759,36743,1736302808837] 2025-01-08T02:24:36,368 DEBUG [RegionServerTracker-0 {}] master.DeadServer(103): Processing c4e5d52b9759,36743,1736302808837; numProcessing=3 2025-01-08T02:24:36,369 DEBUG [RegionServerTracker-0 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/draining/c4e5d52b9759,36743,1736302808837 already deleted, retry=false 2025-01-08T02:24:36,369 INFO [RegionServerTracker-0 {}] master.ServerManager(652): Cluster shutdown set; c4e5d52b9759,36743,1736302808837 expired; onlineServers=0 2025-01-08T02:24:36,369 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2561): ***** STOPPING region server 'c4e5d52b9759,41063,1736302807767' ***** 2025-01-08T02:24:36,369 INFO [RegionServerTracker-0 {}] regionserver.HRegionServer(2575): STOPPED: Cluster shutdown set; onlineServer=0 2025-01-08T02:24:36,369 DEBUG [M:0;c4e5d52b9759:41063 {}] ipc.AbstractRpcClient(198): Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@d844ef7, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=true, bind address=c4e5d52b9759/172.17.0.3:0 2025-01-08T02:24:36,369 INFO [M:0;c4e5d52b9759:41063 {}] regionserver.HRegionServer(1224): stopping server c4e5d52b9759,41063,1736302807767 2025-01-08T02:24:36,369 INFO [M:0;c4e5d52b9759:41063 {}] regionserver.HRegionServer(1250): stopping server c4e5d52b9759,41063,1736302807767; all regions closed. 2025-01-08T02:24:36,369 DEBUG [M:0;c4e5d52b9759:41063 {}] ipc.AbstractRpcClient(514): Stopping rpc client 2025-01-08T02:24:36,369 DEBUG [M:0;c4e5d52b9759:41063 {}] cleaner.LogCleaner(198): Cancelling LogCleaner 2025-01-08T02:24:36,369 DEBUG [M:0;c4e5d52b9759:41063 {}] cleaner.HFileCleaner(335): Stopping file delete threads 2025-01-08T02:24:36,369 WARN [OldWALsCleaner-0 {}] cleaner.LogCleaner(186): Interrupted while cleaning old WALs, will try to clean it next round. Exiting. 2025-01-08T02:24:36,369 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster-HFileCleaner.small.0-1736302810513 {}] cleaner.HFileCleaner(306): Exit Thread[master/c4e5d52b9759:0:becomeActiveMaster-HFileCleaner.small.0-1736302810513,5,FailOnTimeoutGroup] 2025-01-08T02:24:36,369 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster-HFileCleaner.large.0-1736302810509 {}] cleaner.HFileCleaner(306): Exit Thread[master/c4e5d52b9759:0:becomeActiveMaster-HFileCleaner.large.0-1736302810509,5,FailOnTimeoutGroup] 2025-01-08T02:24:36,369 INFO [M:0;c4e5d52b9759:41063 {}] hbase.ChoreService(370): Chore service for: master/c4e5d52b9759:0 had [] on shutdown 2025-01-08T02:24:36,370 DEBUG [M:0;c4e5d52b9759:41063 {}] master.HMaster(1744): Stopping service threads 2025-01-08T02:24:36,370 INFO [M:0;c4e5d52b9759:41063 {}] procedure2.RemoteProcedureDispatcher(119): Stopping procedure remote dispatcher 2025-01-08T02:24:36,370 INFO [M:0;c4e5d52b9759:41063 {}] region.RegionProcedureStore(113): Stopping the Region Procedure Store, isAbort=false 2025-01-08T02:24:36,370 DEBUG [normalizer-worker-0 {}] normalizer.RegionNormalizerWorker(193): interrupt detected. terminating. 2025-01-08T02:24:36,371 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeDeleted, state=SyncConnected, path=/hbase/master 2025-01-08T02:24:36,371 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=NodeChildrenChanged, state=SyncConnected, path=/hbase 2025-01-08T02:24:36,371 DEBUG [M:0;c4e5d52b9759:41063 {}] zookeeper.ZKUtil(347): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Unable to get data of znode /hbase/master because node does not exist (not an error) 2025-01-08T02:24:36,371 WARN [M:0;c4e5d52b9759:41063 {}] master.ActiveMasterManager(344): Failed get of master address: java.io.IOException: Can't get master address from ZooKeeper; znode data == null 2025-01-08T02:24:36,371 DEBUG [zk-event-processor-pool-0 {}] zookeeper.ZKUtil(113): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Set watcher on znode that does not yet exist, /hbase/master 2025-01-08T02:24:36,371 INFO [M:0;c4e5d52b9759:41063 {}] assignment.AssignmentManager(391): Stopping assignment manager 2025-01-08T02:24:36,371 INFO [M:0;c4e5d52b9759:41063 {}] region.MasterRegion(195): Closing local region {ENCODED => 1595e783b53d99cd5eef43b6debb2682, NAME => 'master:store,,1.1595e783b53d99cd5eef43b6debb2682.', STARTKEY => '', ENDKEY => ''}, isAbort=false 2025-01-08T02:24:36,371 DEBUG [M:0;c4e5d52b9759:41063 {}] regionserver.HRegion(1681): Closing 1595e783b53d99cd5eef43b6debb2682, disabling compactions & flushes 2025-01-08T02:24:36,387 INFO [M:0;c4e5d52b9759:41063 {}] regionserver.HRegion(1703): Closing region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T02:24:36,387 DEBUG [M:0;c4e5d52b9759:41063 {}] regionserver.HRegion(1724): Waiting without time limit for close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T02:24:36,387 DEBUG [M:0;c4e5d52b9759:41063 {}] regionserver.HRegion(1791): Acquired close lock on master:store,,1.1595e783b53d99cd5eef43b6debb2682. after waiting 0 ms 2025-01-08T02:24:36,387 DEBUG [M:0;c4e5d52b9759:41063 {}] regionserver.HRegion(1801): Updates disabled for region master:store,,1.1595e783b53d99cd5eef43b6debb2682. 2025-01-08T02:24:36,387 INFO [M:0;c4e5d52b9759:41063 {}] regionserver.HRegion(2837): Flushing 1595e783b53d99cd5eef43b6debb2682 4/4 column families, dataSize=810.93 KB heapSize=972.33 KB 2025-01-08T02:24:36,387 ERROR [AsyncFSWAL-0-hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData-prefix:c4e5d52b9759,41063,1736302807767 {}] server.NIOServerCnxnFactory(85): Thread Thread[AsyncFSWAL-0-hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData-prefix:c4e5d52b9759,41063,1736302807767,5,FailOnTimeoutGroup] died java.lang.NullPointerException: Cannot invoke "org.apache.hbase.thirdparty.io.netty.buffer.ByteBuf.readableBytes()" because "this.buf" is null at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.buffered(FanOutOneBlockAsyncDFSOutput.java:419) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.append(AsyncProtobufLogWriter.java:132) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doAppend(AsyncFSWAL.java:830) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doAppend(AsyncFSWAL.java:128) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.appendEntry(AbstractFSWAL.java:1148) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.appendAndSync(AsyncFSWAL.java:500) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.consume(AsyncFSWAL.java:603) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:24:36,390 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:24:36,464 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:24:36,464 INFO [RS:2;c4e5d52b9759:39911 {}] regionserver.HRegionServer(1307): Exiting; stopping=c4e5d52b9759,39911,1736302808913; zookeeper connection closed. 2025-01-08T02:24:36,464 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:39911-0x1015754588a0003, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:24:36,465 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@38c7f5ed {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@38c7f5ed 2025-01-08T02:24:36,465 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:24:36,465 INFO [RS:0;c4e5d52b9759:37031 {}] regionserver.HRegionServer(1307): Exiting; stopping=c4e5d52b9759,37031,1736302808764; zookeeper connection closed. 2025-01-08T02:24:36,465 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:37031-0x1015754588a0001, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:24:36,465 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@52e30c51 {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@52e30c51 2025-01-08T02:24:36,467 INFO [RS:1;c4e5d52b9759:36743 {}] regionserver.HRegionServer(1307): Exiting; stopping=c4e5d52b9759,36743,1736302808837; zookeeper connection closed. 2025-01-08T02:24:36,467 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:24:36,467 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): regionserver:36743-0x1015754588a0002, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:24:36,467 INFO [Shutdown of org.apache.hadoop.hbase.fs.HFileSystem@41c06aad {}] hbase.MiniHBaseCluster$SingleFileSystemShutdownThread(216): Hook closing fs=org.apache.hadoop.hbase.fs.HFileSystem@41c06aad 2025-01-08T02:24:36,468 INFO [Time-limited test {}] util.JVMClusterUtil(335): Shutdown of 1 master(s) and 3 regionserver(s) complete 2025-01-08T02:24:38,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.RegionServer.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:38,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint 2025-01-08T02:24:38,431 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_meta 2025-01-08T02:24:38,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_default_table_testExportExpiredSnapshot 2025-01-08T02:24:38,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_namespace 2025-01-08T02:24:38,432 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=TableRequests_Namespace_hbase_table_acl 2025-01-08T02:24:38,433 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:38,433 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: Master,sub=Coprocessor.Master.CP_org.apache.hadoop.hbase.security.access.SecureTestUtil$MasterSyncObserver 2025-01-08T02:24:38,433 DEBUG [HBase-Metrics2-1 {}] impl.GlobalMetricRegistriesAdapter(153): Removing adapter for the MetricRegistry: RegionServer,sub=Coprocessor.Region.CP_org.apache.hadoop.hbase.security.access.AccessController 2025-01-08T02:24:41,671 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:25:06,391 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:25:09,041 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): data stats (chunk size=2097152): current pool size=7, created chunk count=10, reused chunk count=22, reuseRatio=68.75% 2025-01-08T02:25:09,042 DEBUG [master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics {}] regionserver.ChunkCreator$MemStoreChunkPool$StatisticsThread(417): index stats (chunk size=209715): current pool size=0, created chunk count=0, reused chunk count=0, reuseRatio=0 2025-01-08T02:25:16,827 WARN [HBase-Metrics2-1 {}] impl.MetricsConfig(138): Cannot locate configuration: tried hadoop-metrics2-jobhistoryserver.properties,hadoop-metrics2.properties 2025-01-08T02:25:36,391 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;c4e5d52b9759:41063 225 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 1 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 32 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@44e8dc0c Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 18 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: WAITING Blocked count: 0 Waited count: 17 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@341aa7cd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: RUNNABLE Blocked count: 0 Waited count: 18 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 3369 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 35 Waiting on java.util.concurrent.CountDownLatch$Sync@5604b40 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12614 Waited count: 13134 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 10 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3f94f9f7 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@5e511507 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@561de66c): State: TIMED_WAITING Blocked count: 0 Waited count: 669 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp1696620586-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp1696620586-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp1696620586-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp1696620586-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp1696620586-41): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp1696620586-42-acceptor-0@60471ea1-ServerConnector@3797254c{HTTP/1.1, (http/1.1)}{localhost:35271}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp1696620586-43): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp1696620586-44): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-70e67bfe-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 26 Waited count: 3377 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51995016 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 43527): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@4db32c52): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 111 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@69e626c6): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 111 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 32924 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 0 Waited count: 1698 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1820a089 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 43527): State: TIMED_WAITING Blocked count: 50 Waited count: 2286 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 43527): State: TIMED_WAITING Blocked count: 82 Waited count: 2304 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 43527): State: TIMED_WAITING Blocked count: 41 Waited count: 2308 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 43527): State: TIMED_WAITING Blocked count: 59 Waited count: 2294 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 43527): State: TIMED_WAITING Blocked count: 40 Waited count: 2304 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7ef1250d): State: TIMED_WAITING Blocked count: 0 Waited count: 167 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@39952a1b): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@1c2660bd): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@3b0f3338): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1523957747)): State: TIMED_WAITING Blocked count: 0 Waited count: 13 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp769527212-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp769527212-87-acceptor-0@3c8530e7-ServerConnector@76ce3a96{HTTP/1.1, (http/1.1)}{localhost:45059}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp769527212-88): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp769527212-89): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-788996d7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@c046f6f): State: TIMED_WAITING Blocked count: 0 Waited count: 664 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 33591): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 0 Waited count: 296 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@16a73765 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1354 Waited count: 1458 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@32e51ee0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 340 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 341 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 347 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 335 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 4 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 338 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp2041437990-121): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp2041437990-122-acceptor-0@241d394-ServerConnector@62d7732{HTTP/1.1, (http/1.1)}{localhost:36645}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp2041437990-123): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (831077991) connection to localhost/127.0.0.1:43527 from jenkins): State: TIMED_WAITING Blocked count: 1416 Waited count: 1417 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 124 (qtp2041437990-124): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-27fafcf5-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 0 Waited count: 2044 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@d3cb706): State: TIMED_WAITING Blocked count: 0 Waited count: 663 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 45047): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 0 Waited count: 275 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7335a455 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1341 Waited count: 1450 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3bfb6ffd): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (IPC Server handler 4 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 332 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1)): State: TIMED_WAITING Blocked count: 13 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 149 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2)): State: TIMED_WAITING Blocked count: 12 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 161 (pool-39-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3)): State: TIMED_WAITING Blocked count: 5 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp85150872-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp85150872-167-acceptor-0@2b65e585-ServerConnector@e141426{HTTP/1.1, (http/1.1)}{localhost:46313}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 168 (qtp85150872-168): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 169 (qtp85150872-169): State: TIMED_WAITING Blocked count: 0 Waited count: 6 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 166 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 171 (Session-HouseKeeper-79cea7f7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 170 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 187 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 189 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (java.util.concurrent.ThreadPoolExecutor$Worker@7c8dd6a7[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (java.util.concurrent.ThreadPoolExecutor$Worker@db1f719[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@2b968e40): State: TIMED_WAITING Blocked count: 0 Waited count: 662 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 36113): State: TIMED_WAITING Blocked count: 1 Waited count: 35 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 67 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 1 Waited count: 270 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@42b32cdf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1334 Waited count: 1455 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@459cf68): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 198 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 331 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 331 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 331 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 334 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 212 (IPC Server handler 4 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 331 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3fdf04ce[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 12 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 14 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:61593): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 34 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 166 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 29 Waited count: 735 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3adc2545 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:61593):): State: WAITING Blocked count: 1 Waited count: 844 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6eb1751a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 877 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@11d419ff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 245 (LeaseRenewer:jenkins@localhost:43527): State: TIMED_WAITING Blocked count: 8 Waited count: 340 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34a48fd7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 253 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 41 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:61593)): State: RUNNABLE Blocked count: 41 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 56 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60f2d5c9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 0 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 37 Waited count: 86 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@44149220 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 5 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 0 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 0 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 130 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@54de7742 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 185 Waited count: 692 Waiting on java.util.concurrent.Semaphore$NonfairSync@44395ff0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 125 Waited count: 495 Waiting on java.util.concurrent.Semaphore$NonfairSync@32939623 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063): State: WAITING Blocked count: 90 Waited count: 4903 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@52f697c3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@7b7d9931 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@55713af3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@33dff711 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@6a467f8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;c4e5d52b9759:41063): State: TIMED_WAITING Blocked count: 6 Waited count: 2675 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$961/0x00007ff380f13618.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 33 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7a779c6a): State: TIMED_WAITING Blocked count: 0 Waited count: 110 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 3267 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 78 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 79 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 54 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 33 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 32588 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@387eb41e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7303e525 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 17 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c180f54 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 18 Waited count: 31 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6dc8d2bc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 493 (LeaseRenewer:jenkins.hfs.1@localhost:43527): State: TIMED_WAITING Blocked count: 8 Waited count: 338 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 505 (LeaseRenewer:jenkins.hfs.0@localhost:43527): State: TIMED_WAITING Blocked count: 10 Waited count: 340 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 508 (LeaseRenewer:jenkins.hfs.2@localhost:43527): State: TIMED_WAITING Blocked count: 8 Waited count: 338 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.run(LeaseRenewer.java:441) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer.access$800(LeaseRenewer.java:77) app//org.apache.hadoop.hdfs.client.impl.LeaseRenewer$1.run(LeaseRenewer.java:336) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 9 Waited count: 15 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 554 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 32367 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 583 (ForkJoinPool.commonPool-worker-2): State: WAITING Blocked count: 0 Waited count: 327 Waiting on java.util.concurrent.ForkJoinPool@4c9db6ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 584 (ForkJoinPool.commonPool-worker-3): State: WAITING Blocked count: 0 Waited count: 497 Waiting on java.util.concurrent.ForkJoinPool@4c9db6ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 594 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 1 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 2 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 398 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1109 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 60 Waited count: 96 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5bb51352 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1566 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@213e44c6 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2096 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2998 (region-location-4): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5090 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5091 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5092 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5515 (ForkJoinPool.commonPool-worker-4): State: TIMED_WAITING Blocked count: 0 Waited count: 408 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 9177 (AsyncFSWAL-1-hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData-prefix:c4e5d52b9759,41063,1736302807767): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@37ae9b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9181 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T02:26:06,391 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:26:36,391 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;c4e5d52b9759:41063 220 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 1 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 32 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@44e8dc0c Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 19 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: WAITING Blocked count: 0 Waited count: 20 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@341aa7cd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: RUNNABLE Blocked count: 0 Waited count: 21 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 3969 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 41 Waiting on java.util.concurrent.CountDownLatch$Sync@185588ab Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12614 Waited count: 13135 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 10 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3f94f9f7 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@5e511507 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@561de66c): State: TIMED_WAITING Blocked count: 0 Waited count: 789 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp1696620586-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp1696620586-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp1696620586-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp1696620586-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp1696620586-41): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp1696620586-42-acceptor-0@60471ea1-ServerConnector@3797254c{HTTP/1.1, (http/1.1)}{localhost:35271}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp1696620586-43): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp1696620586-44): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-70e67bfe-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 26 Waited count: 3377 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51995016 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 43527): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@4db32c52): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 131 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@69e626c6): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 131 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 38889 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 0 Waited count: 1698 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1820a089 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 43527): State: TIMED_WAITING Blocked count: 50 Waited count: 2346 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 43527): State: TIMED_WAITING Blocked count: 82 Waited count: 2365 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 43527): State: TIMED_WAITING Blocked count: 41 Waited count: 2369 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 43527): State: TIMED_WAITING Blocked count: 59 Waited count: 2355 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 43527): State: TIMED_WAITING Blocked count: 40 Waited count: 2365 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7ef1250d): State: TIMED_WAITING Blocked count: 0 Waited count: 197 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@39952a1b): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@1c2660bd): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@3b0f3338): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1523957747)): State: TIMED_WAITING Blocked count: 0 Waited count: 15 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp769527212-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp769527212-87-acceptor-0@3c8530e7-ServerConnector@76ce3a96{HTTP/1.1, (http/1.1)}{localhost:45059}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp769527212-88): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp769527212-89): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-788996d7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@c046f6f): State: TIMED_WAITING Blocked count: 0 Waited count: 784 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 33591): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 0 Waited count: 316 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@16a73765 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1374 Waited count: 1498 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@32e51ee0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 405 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 401 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 409 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 395 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 4 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 403 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp2041437990-121): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp2041437990-122-acceptor-0@241d394-ServerConnector@62d7732{HTTP/1.1, (http/1.1)}{localhost:36645}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp2041437990-123): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (831077991) connection to localhost/127.0.0.1:43527 from jenkins): State: TIMED_WAITING Blocked count: 1476 Waited count: 1477 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 124 (qtp2041437990-124): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-27fafcf5-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 0 Waited count: 2104 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@d3cb706): State: TIMED_WAITING Blocked count: 0 Waited count: 783 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 45047): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 0 Waited count: 295 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7335a455 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1361 Waited count: 1490 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3bfb6ffd): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (IPC Server handler 4 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 392 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1)): State: TIMED_WAITING Blocked count: 13 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 149 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2)): State: TIMED_WAITING Blocked count: 12 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 161 (pool-39-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3)): State: TIMED_WAITING Blocked count: 5 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp85150872-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp85150872-167-acceptor-0@2b65e585-ServerConnector@e141426{HTTP/1.1, (http/1.1)}{localhost:46313}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 168 (qtp85150872-168): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 169 (qtp85150872-169): State: TIMED_WAITING Blocked count: 0 Waited count: 7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 166 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 171 (Session-HouseKeeper-79cea7f7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 170 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 187 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 189 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (java.util.concurrent.ThreadPoolExecutor$Worker@7c8dd6a7[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (java.util.concurrent.ThreadPoolExecutor$Worker@db1f719[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@2b968e40): State: TIMED_WAITING Blocked count: 0 Waited count: 782 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 36113): State: TIMED_WAITING Blocked count: 1 Waited count: 41 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 79 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 1 Waited count: 290 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@42b32cdf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1354 Waited count: 1495 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@459cf68): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 198 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 391 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 391 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 391 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 394 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 212 (IPC Server handler 4 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 391 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3fdf04ce[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 14 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 14 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:61593): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 40 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 196 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 29 Waited count: 740 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3adc2545 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:61593):): State: WAITING Blocked count: 1 Waited count: 849 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6eb1751a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 882 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@11d419ff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34a48fd7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 281 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 41 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:61593)): State: RUNNABLE Blocked count: 41 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 56 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60f2d5c9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 37 Waited count: 86 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@44149220 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 5 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 0 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 131 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@54de7742 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 185 Waited count: 692 Waiting on java.util.concurrent.Semaphore$NonfairSync@44395ff0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 125 Waited count: 495 Waiting on java.util.concurrent.Semaphore$NonfairSync@32939623 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063): State: WAITING Blocked count: 90 Waited count: 4903 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@52f697c3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@7b7d9931 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@55713af3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@33dff711 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@6a467f8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;c4e5d52b9759:41063): State: TIMED_WAITING Blocked count: 6 Waited count: 2675 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$961/0x00007ff380f13618.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 39 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7a779c6a): State: TIMED_WAITING Blocked count: 0 Waited count: 130 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 3867 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 78 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 79 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34577861 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 39 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 38590 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@387eb41e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7303e525 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 17 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c180f54 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 18 Waited count: 31 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6dc8d2bc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 9 Waited count: 15 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 554 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 38369 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 583 (ForkJoinPool.commonPool-worker-2): State: WAITING Blocked count: 0 Waited count: 327 Waiting on java.util.concurrent.ForkJoinPool@4c9db6ac Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1724) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 584 (ForkJoinPool.commonPool-worker-3): State: TIMED_WAITING Blocked count: 0 Waited count: 498 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 594 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 1 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 2 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 404 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1109 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 60 Waited count: 96 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5bb51352 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1566 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@213e44c6 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2096 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2998 (region-location-4): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5090 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5091 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5092 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9177 (AsyncFSWAL-1-hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData-prefix:c4e5d52b9759,41063,1736302807767): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@37ae9b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9181 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T02:27:06,391 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:27:36,392 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;c4e5d52b9759:41063 219 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 1 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 32 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@44e8dc0c Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 20 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: WAITING Blocked count: 0 Waited count: 23 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@341aa7cd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: RUNNABLE Blocked count: 0 Waited count: 24 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 4569 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 47 Waiting on java.util.concurrent.CountDownLatch$Sync@4b1ee713 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12614 Waited count: 13136 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 10 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3f94f9f7 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@5e511507 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@561de66c): State: TIMED_WAITING Blocked count: 0 Waited count: 909 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp1696620586-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp1696620586-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp1696620586-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp1696620586-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp1696620586-41): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp1696620586-42-acceptor-0@60471ea1-ServerConnector@3797254c{HTTP/1.1, (http/1.1)}{localhost:35271}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp1696620586-43): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp1696620586-44): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-70e67bfe-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 26 Waited count: 3377 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51995016 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 43527): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@4db32c52): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 151 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@69e626c6): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 151 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 44853 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 0 Waited count: 1698 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1820a089 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 43527): State: TIMED_WAITING Blocked count: 50 Waited count: 2408 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 43527): State: TIMED_WAITING Blocked count: 82 Waited count: 2426 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 43527): State: TIMED_WAITING Blocked count: 41 Waited count: 2430 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 43527): State: TIMED_WAITING Blocked count: 59 Waited count: 2415 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 43527): State: TIMED_WAITING Blocked count: 40 Waited count: 2426 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7ef1250d): State: TIMED_WAITING Blocked count: 0 Waited count: 227 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@39952a1b): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@1c2660bd): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@3b0f3338): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1523957747)): State: TIMED_WAITING Blocked count: 0 Waited count: 17 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp769527212-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp769527212-87-acceptor-0@3c8530e7-ServerConnector@76ce3a96{HTTP/1.1, (http/1.1)}{localhost:45059}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp769527212-88): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp769527212-89): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-788996d7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@c046f6f): State: TIMED_WAITING Blocked count: 0 Waited count: 904 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 33591): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 0 Waited count: 336 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@16a73765 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1394 Waited count: 1538 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@32e51ee0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 465 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 461 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 470 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 455 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 4 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 467 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp2041437990-121): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp2041437990-122-acceptor-0@241d394-ServerConnector@62d7732{HTTP/1.1, (http/1.1)}{localhost:36645}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp2041437990-123): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (831077991) connection to localhost/127.0.0.1:43527 from jenkins): State: TIMED_WAITING Blocked count: 1536 Waited count: 1537 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 124 (qtp2041437990-124): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-27fafcf5-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 0 Waited count: 2164 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@d3cb706): State: TIMED_WAITING Blocked count: 0 Waited count: 903 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 45047): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 0 Waited count: 315 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7335a455 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1381 Waited count: 1530 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3bfb6ffd): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (IPC Server handler 4 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 452 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1)): State: TIMED_WAITING Blocked count: 13 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 149 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2)): State: TIMED_WAITING Blocked count: 12 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 161 (pool-39-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3)): State: TIMED_WAITING Blocked count: 5 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp85150872-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp85150872-167-acceptor-0@2b65e585-ServerConnector@e141426{HTTP/1.1, (http/1.1)}{localhost:46313}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 168 (qtp85150872-168): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 169 (qtp85150872-169): State: TIMED_WAITING Blocked count: 0 Waited count: 8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 166 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 171 (Session-HouseKeeper-79cea7f7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 170 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 187 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 189 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (java.util.concurrent.ThreadPoolExecutor$Worker@7c8dd6a7[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (java.util.concurrent.ThreadPoolExecutor$Worker@db1f719[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@2b968e40): State: TIMED_WAITING Blocked count: 0 Waited count: 902 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 36113): State: TIMED_WAITING Blocked count: 1 Waited count: 47 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 91 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 1 Waited count: 310 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@42b32cdf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1374 Waited count: 1535 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@459cf68): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 198 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 451 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 451 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 451 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 454 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 212 (IPC Server handler 4 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 451 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3fdf04ce[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 16 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 14 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:61593): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 46 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 226 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 29 Waited count: 744 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3adc2545 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:61593):): State: WAITING Blocked count: 1 Waited count: 853 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6eb1751a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 886 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@11d419ff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34a48fd7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 309 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 41 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:61593)): State: RUNNABLE Blocked count: 41 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 56 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60f2d5c9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 37 Waited count: 86 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@44149220 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 5 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 0 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@54de7742 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 185 Waited count: 692 Waiting on java.util.concurrent.Semaphore$NonfairSync@44395ff0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 125 Waited count: 495 Waiting on java.util.concurrent.Semaphore$NonfairSync@32939623 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063): State: WAITING Blocked count: 90 Waited count: 4903 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@52f697c3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@7b7d9931 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@55713af3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@33dff711 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@6a467f8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;c4e5d52b9759:41063): State: TIMED_WAITING Blocked count: 6 Waited count: 2675 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$961/0x00007ff380f13618.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 45 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7a779c6a): State: TIMED_WAITING Blocked count: 0 Waited count: 150 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 4467 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 78 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 79 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34577861 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 45 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 44591 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@387eb41e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7303e525 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 17 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c180f54 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 18 Waited count: 31 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6dc8d2bc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 9 Waited count: 15 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 554 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 44371 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 583 (ForkJoinPool.commonPool-worker-2): State: TIMED_WAITING Blocked count: 0 Waited count: 328 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkUntil(LockSupport.java:410) java.base@17.0.11/java.util.concurrent.ForkJoinPool.awaitWork(ForkJoinPool.java:1726) java.base@17.0.11/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1623) java.base@17.0.11/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:165) Thread 594 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 1 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 2 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 410 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1109 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 60 Waited count: 96 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5bb51352 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1566 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@213e44c6 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2096 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2998 (region-location-4): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5090 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5091 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5092 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9177 (AsyncFSWAL-1-hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData-prefix:c4e5d52b9759,41063,1736302807767): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@37ae9b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9181 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 14 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T02:28:06,392 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:28:36,392 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;c4e5d52b9759:41063 218 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 1 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 32 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@44e8dc0c Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 21 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: WAITING Blocked count: 0 Waited count: 26 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@341aa7cd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: RUNNABLE Blocked count: 0 Waited count: 27 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 5168 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 53 Waiting on java.util.concurrent.CountDownLatch$Sync@7660db9d Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12614 Waited count: 13137 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 10 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3f94f9f7 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@5e511507 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@561de66c): State: TIMED_WAITING Blocked count: 0 Waited count: 1029 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp1696620586-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp1696620586-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp1696620586-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp1696620586-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp1696620586-41): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp1696620586-42-acceptor-0@60471ea1-ServerConnector@3797254c{HTTP/1.1, (http/1.1)}{localhost:35271}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp1696620586-43): State: TIMED_WAITING Blocked count: 0 Waited count: 13 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp1696620586-44): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-70e67bfe-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 26 Waited count: 3377 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51995016 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 43527): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@4db32c52): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 171 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@69e626c6): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 171 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 50818 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 0 Waited count: 1698 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1820a089 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 43527): State: TIMED_WAITING Blocked count: 50 Waited count: 2469 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 43527): State: TIMED_WAITING Blocked count: 82 Waited count: 2487 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 43527): State: TIMED_WAITING Blocked count: 41 Waited count: 2490 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 43527): State: TIMED_WAITING Blocked count: 59 Waited count: 2476 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 43527): State: TIMED_WAITING Blocked count: 40 Waited count: 2486 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7ef1250d): State: TIMED_WAITING Blocked count: 0 Waited count: 257 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@39952a1b): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@1c2660bd): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@3b0f3338): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1523957747)): State: TIMED_WAITING Blocked count: 0 Waited count: 19 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp769527212-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp769527212-87-acceptor-0@3c8530e7-ServerConnector@76ce3a96{HTTP/1.1, (http/1.1)}{localhost:45059}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp769527212-88): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp769527212-89): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-788996d7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@c046f6f): State: TIMED_WAITING Blocked count: 0 Waited count: 1024 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 33591): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 0 Waited count: 356 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@16a73765 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1414 Waited count: 1578 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@32e51ee0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 525 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 521 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 531 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 515 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 4 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 527 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp2041437990-121): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp2041437990-122-acceptor-0@241d394-ServerConnector@62d7732{HTTP/1.1, (http/1.1)}{localhost:36645}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp2041437990-123): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (831077991) connection to localhost/127.0.0.1:43527 from jenkins): State: TIMED_WAITING Blocked count: 1596 Waited count: 1597 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 124 (qtp2041437990-124): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-27fafcf5-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 0 Waited count: 2224 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@d3cb706): State: TIMED_WAITING Blocked count: 0 Waited count: 1023 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 45047): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 0 Waited count: 335 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7335a455 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1401 Waited count: 1570 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3bfb6ffd): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (IPC Server handler 4 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 512 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1)): State: TIMED_WAITING Blocked count: 13 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 149 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2)): State: TIMED_WAITING Blocked count: 12 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 161 (pool-39-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3)): State: TIMED_WAITING Blocked count: 5 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp85150872-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp85150872-167-acceptor-0@2b65e585-ServerConnector@e141426{HTTP/1.1, (http/1.1)}{localhost:46313}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 168 (qtp85150872-168): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 169 (qtp85150872-169): State: TIMED_WAITING Blocked count: 0 Waited count: 9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 166 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 171 (Session-HouseKeeper-79cea7f7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 170 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 187 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 189 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (java.util.concurrent.ThreadPoolExecutor$Worker@7c8dd6a7[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (java.util.concurrent.ThreadPoolExecutor$Worker@db1f719[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@2b968e40): State: TIMED_WAITING Blocked count: 0 Waited count: 1022 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 36113): State: TIMED_WAITING Blocked count: 1 Waited count: 53 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 103 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 1 Waited count: 330 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@42b32cdf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1394 Waited count: 1575 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@459cf68): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 198 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 511 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 511 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 511 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 514 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 212 (IPC Server handler 4 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 511 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3fdf04ce[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 18 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 14 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:61593): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 52 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 256 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 29 Waited count: 749 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3adc2545 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:61593):): State: WAITING Blocked count: 1 Waited count: 858 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6eb1751a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 891 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@11d419ff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34a48fd7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 337 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 41 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:61593)): State: RUNNABLE Blocked count: 41 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 56 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60f2d5c9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 37 Waited count: 86 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@44149220 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 5 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 132 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@54de7742 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 185 Waited count: 692 Waiting on java.util.concurrent.Semaphore$NonfairSync@44395ff0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 125 Waited count: 495 Waiting on java.util.concurrent.Semaphore$NonfairSync@32939623 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063): State: WAITING Blocked count: 90 Waited count: 4903 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@52f697c3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@7b7d9931 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@55713af3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@33dff711 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@6a467f8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;c4e5d52b9759:41063): State: TIMED_WAITING Blocked count: 6 Waited count: 2675 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:169) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$$Lambda$961/0x00007ff380f13618.run(Unknown Source) app//org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) app//org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) app//org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) app//org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) app//org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) app//org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 51 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7a779c6a): State: TIMED_WAITING Blocked count: 0 Waited count: 170 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 5066 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 78 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 79 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34577861 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 51 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 50593 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@387eb41e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7303e525 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 17 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c180f54 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 18 Waited count: 31 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6dc8d2bc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 9 Waited count: 15 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 554 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 50373 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 594 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 1 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 2 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 416 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1109 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 60 Waited count: 96 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5bb51352 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1566 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@213e44c6 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2096 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2998 (region-location-4): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5090 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5091 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5092 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9177 (AsyncFSWAL-1-hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData-prefix:c4e5d52b9759,41063,1736302807767): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@37ae9b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9181 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 20 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) 2025-01-08T02:29:06,392 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:29:36,388 DEBUG [M:0;c4e5d52b9759:41063 {}] regionserver.HRegion(1635): Region close journal for 1595e783b53d99cd5eef43b6debb2682: 2025-01-08T02:29:36,388 WARN [M:0;c4e5d52b9759:41063 {}] region.MasterRegion(134): Failed to close region org.apache.hadoop.hbase.regionserver.wal.WALSyncTimeoutIOException: org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync result after 300000 ms for txid=3722, WAL system stuck? at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:883) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doSync(AsyncFSWAL.java:671) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.lambda$sync$1(AbstractFSWAL.java:600) ~[classes/:?] at org.apache.hadoop.hbase.trace.TraceUtil.trace(TraceUtil.java:187) ~[hbase-common-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:600) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.sync(AbstractFSWAL.java:590) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.doSyncOfUnflushedWALChanges(HRegion.java:2869) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalPrepareFlushCache(HRegion.java:2811) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2670) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2644) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.internalFlushcache(HRegion.java:2635) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.doClose(HRegion.java:1810) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1631) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1586) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegion.close(HRegion.java:1569) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.closeRegion(MasterRegion.java:132) ~[classes/:?] at org.apache.hadoop.hbase.master.region.MasterRegion.close(MasterRegion.java:205) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.stopServiceThreads(HMaster.java:1769) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1285) ~[classes/:?] at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:614) ~[classes/:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: org.apache.hadoop.hbase.exceptions.TimeoutIOException: Failed to get sync result after 300000 ms for txid=3722, WAL system stuck? at org.apache.hadoop.hbase.regionserver.wal.SyncFuture.get(SyncFuture.java:171) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.blockOnSync(AbstractFSWAL.java:879) ~[classes/:?] ... 20 more 2025-01-08T02:29:36,396 DEBUG [FsDatasetAsyncDiskServiceFixer {}] hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer(620): NoSuchFieldException: threadGroup; It might because your Hadoop version > 3.2.3 or 3.3.4, See HBASE-27595 for details. 2025-01-08T02:29:36,396 WARN [Close-WAL-Writer-0 {}] wal.AsyncProtobufLogWriter(163): normal close failed, try recover java.lang.NullPointerException: Cannot invoke "org.apache.hbase.thirdparty.io.netty.buffer.ByteBuf.ensureWritable(int)" because "this.buf" is null at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.writeInt(FanOutOneBlockAsyncDFSOutput.java:396) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.lambda$writeWALTrailerAndMagic$3(AsyncProtobufLogWriter.java:243) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.writeWALMetadata(AsyncProtobufLogWriter.java:201) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.writeWALTrailerAndMagic(AsyncProtobufLogWriter.java:236) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AbstractProtobufLogWriter.writeWALTrailer(AbstractProtobufLogWriter.java:252) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:160) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:29:36,398 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(47): Initialize RecoverLeaseFSUtils 2025-01-08T02:29:36,398 DEBUG [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(59): set recoverLeaseMethod to org.apache.hadoop.fs.LeaseRecoverable.recoverLease() 2025-01-08T02:29:36,398 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file /user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644 2025-01-08T02:29:36,399 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=0 on file=/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644 after 1ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.recoverAndClose(FanOutOneBlockAsyncDFSOutput.java:610) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:164) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:29:36,399 WARN [Close-WAL-Writer-0 {}] wal.AsyncFSWAL(734): close old writer failed. java.io.InterruptedIOException: Operation cancelled at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.checkIfCancelled(RecoverLeaseFSUtils.java:269) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:159) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.io.asyncfs.FanOutOneBlockAsyncDFSOutput.recoverAndClose(FanOutOneBlockAsyncDFSOutput.java:610) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:2.7.0-SNAPSHOT] at org.apache.hadoop.hbase.regionserver.wal.AsyncProtobufLogWriter.close(AsyncProtobufLogWriter.java:164) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:732) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:29:36,399 INFO [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(134): Recover lease on dfs file hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644 2025-01-08T02:29:36,399 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=0 on file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644 after 0ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Process Thread Dump: Automatic Stack Trace every 60 seconds waiting on M:0;c4e5d52b9759:41063 221 active threads Thread 1 (main): State: TIMED_WAITING Blocked count: 1 Waited count: 4 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.junit.internal.runners.statements.FailOnTimeout.getResult(FailOnTimeout.java:167) app//org.junit.internal.runners.statements.FailOnTimeout.evaluate(FailOnTimeout.java:128) app//org.apache.hadoop.hbase.SystemExitRule$1.evaluate(SystemExitRule.java:39) app//org.junit.rules.RunRules.evaluate(RunRules.java:20) app//org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) app//org.junit.runners.ParentRunner.run(ParentRunner.java:413) app//org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:316) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:240) app//org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:214) app//org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:155) app//org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:385) app//org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) app//org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:507) app//org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:495) Thread 2 (Reference Handler): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/java.lang.ref.Reference.waitForReferencePendingList(Native Method) java.base@17.0.11/java.lang.ref.Reference.processPendingReferences(Reference.java:253) java.base@17.0.11/java.lang.ref.Reference$ReferenceHandler.run(Reference.java:215) Thread 3 (Finalizer): State: WAITING Blocked count: 32 Waited count: 13 Waiting on java.lang.ref.ReferenceQueue$Lock@44e8dc0c Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) java.base@17.0.11/java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:172) Thread 4 (Signal Dispatcher): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 12 (Common-Cleaner): State: TIMED_WAITING Blocked count: 18 Waited count: 22 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/jdk.internal.ref.CleanerImpl.run(CleanerImpl.java:140) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) java.base@17.0.11/jdk.internal.misc.InnocuousThread.run(InnocuousThread.java:162) Thread 13 (Notification Thread): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: Thread 14 (pool-1-thread-1): State: WAITING Blocked count: 0 Waited count: 28 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@341aa7cd Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:275) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 15 (pool-1-thread-2): State: RUNNABLE Blocked count: 0 Waited count: 31 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.poll(EPollPort.java:200) java.base@17.0.11/sun.nio.ch.EPollPort$EventHandlerTask.run(EPollPort.java:281) java.base@17.0.11/sun.nio.ch.AsynchronousChannelGroupImpl$1.run(AsynchronousChannelGroupImpl.java:113) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 16 (surefire-forkedjvm-stream-flusher): State: TIMED_WAITING Blocked count: 0 Waited count: 5768 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 18 (surefire-forkedjvm-command-thread): State: WAITING Blocked count: 0 Waited count: 59 Waiting on java.util.concurrent.CountDownLatch$Sync@540365bb Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.CountDownLatch.await(CountDownLatch.java:230) java.base@17.0.11/sun.nio.ch.PendingFuture.get(PendingFuture.java:178) app//org.apache.maven.surefire.api.util.internal.Channels$2.read(Channels.java:127) java.base@17.0.11/java.io.BufferedInputStream.fill(BufferedInputStream.java:244) java.base@17.0.11/java.io.BufferedInputStream.read1(BufferedInputStream.java:284) java.base@17.0.11/java.io.BufferedInputStream.read(BufferedInputStream.java:343) app//org.apache.maven.surefire.api.util.internal.Channels$3.readImpl(Channels.java:169) app//org.apache.maven.surefire.api.util.internal.AbstractNoninterruptibleReadableChannel.read(AbstractNoninterruptibleReadableChannel.java:50) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:430) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.read(AbstractStreamDecoder.java:419) app//org.apache.maven.surefire.api.stream.AbstractStreamDecoder.readMessageType(AbstractStreamDecoder.java:116) app//org.apache.maven.surefire.booter.stream.CommandDecoder.decode(CommandDecoder.java:77) app//org.apache.maven.surefire.booter.spi.CommandChannelDecoder.decode(CommandChannelDecoder.java:60) app//org.apache.maven.surefire.booter.CommandReader$CommandRunnable.run(CommandReader.java:290) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 22 (Time-limited test): State: RUNNABLE Blocked count: 12614 Waited count: 13138 Stack: java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo1(Native Method) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:197) java.management@17.0.11/sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:154) app//org.apache.hadoop.hbase.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:181) app//org.apache.hadoop.hbase.util.Threads.printThreadInfo(Threads.java:186) app//org.apache.hadoop.hbase.util.Threads.threadDumpingIsAlive(Threads.java:113) app//org.apache.hadoop.hbase.LocalHBaseCluster.join(LocalHBaseCluster.java:394) app//org.apache.hadoop.hbase.MiniHBaseCluster.waitUntilShutDown(MiniHBaseCluster.java:921) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniHBaseCluster(HBaseTestingUtility.java:1359) app//org.apache.hadoop.hbase.HBaseTestingUtility.shutdownMiniCluster(HBaseTestingUtility.java:1341) app//org.apache.hadoop.hbase.snapshot.TestExportSnapshot.tearDownAfterClass(TestExportSnapshot.java:121) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) java.base@17.0.11/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) java.base@17.0.11/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.base@17.0.11/java.lang.reflect.Method.invoke(Method.java:568) app//org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) app//org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) app//org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) app//org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) app//org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) Thread 23 (org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner): State: WAITING Blocked count: 10 Waited count: 10 Waiting on java.lang.ref.ReferenceQueue$Lock@3f94f9f7 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:155) java.base@17.0.11/java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:176) app//org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner.run(FileSystem.java:4171) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 25 (SSL Certificates Store Monitor): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.TaskQueue@5e511507 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 34 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@561de66c): State: TIMED_WAITING Blocked count: 0 Waited count: 1149 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 35 (GcTimeMonitor obsWindow = 60000, sleepInterval = 5000, maxGcTimePerc = 100): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.GcTimeMonitor.run(GcTimeMonitor.java:161) Thread 36 (pool-6-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 37 (qtp1696620586-37): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 38 (qtp1696620586-38): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 39 (qtp1696620586-39): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 40 (qtp1696620586-40): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 41 (qtp1696620586-41): State: TIMED_WAITING Blocked count: 0 Waited count: 11 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 42 (qtp1696620586-42-acceptor-0@60471ea1-ServerConnector@3797254c{HTTP/1.1, (http/1.1)}{localhost:35271}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 43 (qtp1696620586-43): State: TIMED_WAITING Blocked count: 0 Waited count: 14 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 44 (qtp1696620586-44): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 45 (Session-HouseKeeper-70e67bfe-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 46 (pool-7-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 53 (FSEditLogAsync): State: WAITING Blocked count: 26 Waited count: 3377 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@51995016 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.dequeueEdit(FSEditLogAsync.java:241) app//org.apache.hadoop.hdfs.server.namenode.FSEditLogAsync.run(FSEditLogAsync.java:250) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 55 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 56 (IPC Server idle connection scanner for port 43527): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 58 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 61 (org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor@4db32c52): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.PendingReconstructionBlocks$PendingReconstructionMonitor.run(PendingReconstructionBlocks.java:267) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 62 (DatanodeAdminMonitor-0): State: TIMED_WAITING Blocked count: 0 Waited count: 191 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 49 (org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor@69e626c6): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.HeartbeatManager$Monitor.run(HeartbeatManager.java:563) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 47 (RedundancyMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 191 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) java.base@17.0.11/java.lang.Thread.sleep(Thread.java:344) java.base@17.0.11/java.util.concurrent.TimeUnit.sleep(TimeUnit.java:446) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$RedundancyMonitor.run(BlockManager.java:5352) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 48 (MarkedDeleteBlockScrubberThread): State: TIMED_WAITING Blocked count: 0 Waited count: 56775 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$MarkedDeleteBlockScrubber.run(BlockManager.java:5326) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 51 (Block report processor): State: WAITING Blocked count: 0 Waited count: 1698 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@1820a089 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ArrayBlockingQueue.take(ArrayBlockingQueue.java:420) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.processQueue(BlockManager.java:5627) app//org.apache.hadoop.hdfs.server.blockmanagement.BlockManager$BlockReportProcessingThread.run(BlockManager.java:5614) Thread 57 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 54 (IPC Server listener on 0): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 64 (IPC Server handler 0 on default port 43527): State: TIMED_WAITING Blocked count: 50 Waited count: 2530 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 65 (IPC Server handler 1 on default port 43527): State: TIMED_WAITING Blocked count: 82 Waited count: 2547 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 66 (IPC Server handler 2 on default port 43527): State: TIMED_WAITING Blocked count: 41 Waited count: 2551 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 67 (IPC Server handler 3 on default port 43527): State: TIMED_WAITING Blocked count: 59 Waited count: 2537 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 68 (IPC Server handler 4 on default port 43527): State: TIMED_WAITING Blocked count: 40 Waited count: 2547 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 69 (pool-12-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 71 (org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor@7ef1250d): State: TIMED_WAITING Blocked count: 0 Waited count: 287 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.LeaseManager$Monitor.run(LeaseManager.java:537) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 72 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor@39952a1b): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeResourceMonitor.run(FSNamesystem.java:4550) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 73 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller@1c2660bd): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$NameNodeEditLogRoller.run(FSNamesystem.java:4592) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 74 (org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber@3b0f3338): State: TIMED_WAITING Blocked count: 0 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.server.namenode.FSNamesystem$LazyPersistFileScrubber.run(FSNamesystem.java:4689) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 75 (CacheReplicationMonitor(1523957747)): State: TIMED_WAITING Blocked count: 0 Waited count: 21 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1759) app//org.apache.hadoop.hdfs.server.blockmanagement.CacheReplicationMonitor.run(CacheReplicationMonitor.java:186) Thread 85 (pool-18-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 86 (qtp769527212-86): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 87 (qtp769527212-87-acceptor-0@3c8530e7-ServerConnector@76ce3a96{HTTP/1.1, (http/1.1)}{localhost:45059}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 88 (qtp769527212-88): State: TIMED_WAITING Blocked count: 0 Waited count: 12 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 89 (qtp769527212-89): State: TIMED_WAITING Blocked count: 0 Waited count: 12 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 90 (Session-HouseKeeper-788996d7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 91 (nioEventLoopGroup-2-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 92 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@c046f6f): State: TIMED_WAITING Blocked count: 0 Waited count: 1144 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 94 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 95 (IPC Server idle connection scanner for port 33591): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 97 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 100 (Command processor): State: WAITING Blocked count: 0 Waited count: 376 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@16a73765 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 101 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1434 Waited count: 1618 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 102 (pool-20-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 84 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@32e51ee0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 96 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 93 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 103 (IPC Server handler 0 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 585 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 104 (IPC Server handler 1 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 583 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 105 (IPC Server handler 2 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 591 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 106 (IPC Server handler 3 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 575 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 107 (IPC Server handler 4 on default port 33591): State: TIMED_WAITING Blocked count: 0 Waited count: 587 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 120 (pool-26-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 121 (qtp2041437990-121): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 122 (qtp2041437990-122-acceptor-0@241d394-ServerConnector@62d7732{HTTP/1.1, (http/1.1)}{localhost:36645}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 123 (qtp2041437990-123): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 118 (IPC Client (831077991) connection to localhost/127.0.0.1:43527 from jenkins): State: TIMED_WAITING Blocked count: 1656 Waited count: 1657 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.ipc.Client$Connection.waitForWork(Client.java:1042) app//org.apache.hadoop.ipc.Client$Connection.run(Client.java:1093) Thread 124 (qtp2041437990-124): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 125 (Session-HouseKeeper-27fafcf5-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 119 (IPC Parameter Sending Thread for localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 0 Waited count: 2284 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferQueue.transfer(SynchronousQueue.java:704) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) app//org.apache.hadoop.ipc.Client$Connection$RpcRequestSender.run(Client.java:1121) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 126 (nioEventLoopGroup-4-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 127 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@d3cb706): State: TIMED_WAITING Blocked count: 0 Waited count: 1143 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 129 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 130 (IPC Server idle connection scanner for port 45047): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 133 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 137 (Command processor): State: WAITING Blocked count: 0 Waited count: 355 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7335a455 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 138 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1421 Waited count: 1610 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 139 (pool-29-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 117 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@3bfb6ffd): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 132 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 128 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 140 (IPC Server handler 0 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 141 (IPC Server handler 1 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 142 (IPC Server handler 2 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 143 (IPC Server handler 3 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 145 (IPC Server handler 4 on default port 45047): State: TIMED_WAITING Blocked count: 0 Waited count: 572 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 146 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1)): State: TIMED_WAITING Blocked count: 13 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 149 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2)): State: TIMED_WAITING Blocked count: 12 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 161 (pool-39-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 162 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 163 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3)): State: TIMED_WAITING Blocked count: 5 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 164 (qtp85150872-164): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.eclipse.jetty.io.ManagedSelector.nioSelect(ManagedSelector.java:183) app//org.eclipse.jetty.io.ManagedSelector.select(ManagedSelector.java:190) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.select(ManagedSelector.java:606) app//org.eclipse.jetty.io.ManagedSelector$SelectorProducer.produce(ManagedSelector.java:543) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produceTask(EatWhatYouKill.java:362) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:186) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173) app//org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.produce(EatWhatYouKill.java:137) app//org.eclipse.jetty.io.ManagedSelector$$Lambda$255/0x00007ff380429138.run(Unknown Source) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 167 (qtp85150872-167-acceptor-0@2b65e585-ServerConnector@e141426{HTTP/1.1, (http/1.1)}{localhost:46313}): State: RUNNABLE Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) app//org.eclipse.jetty.server.ServerConnector.accept(ServerConnector.java:388) app//org.eclipse.jetty.server.AbstractConnector$Acceptor.run(AbstractConnector.java:704) app//org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 168 (qtp85150872-168): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 169 (qtp85150872-169): State: TIMED_WAITING Blocked count: 0 Waited count: 10 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) app//org.eclipse.jetty.util.BlockingArrayQueue.poll(BlockingArrayQueue.java:382) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.idleJobPoll(QueuedThreadPool.java:974) app//org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1018) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 166 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4)): State: TIMED_WAITING Blocked count: 4 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 171 (Session-HouseKeeper-79cea7f7-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 170 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 176 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 177 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 187 (pool-15-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 189 (pool-23-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 195 (java.util.concurrent.ThreadPoolExecutor$Worker@7c8dd6a7[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 192 (java.util.concurrent.ThreadPoolExecutor$Worker@db1f719[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 196 (nioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:994) app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 197 (org.apache.hadoop.util.JvmPauseMonitor$Monitor@2b968e40): State: TIMED_WAITING Blocked count: 0 Waited count: 1142 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.util.JvmPauseMonitor$Monitor.run(JvmPauseMonitor.java:189) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 199 (Socket Reader #1 for port 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener$Reader.doRunLoop(Server.java:1497) app//org.apache.hadoop.ipc.Server$Listener$Reader.run(Server.java:1476) Thread 200 (IPC Server idle connection scanner for port 36113): State: TIMED_WAITING Blocked count: 1 Waited count: 59 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 202 (Hadoop-Metrics-Updater-0): State: TIMED_WAITING Blocked count: 0 Waited count: 115 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 205 (Command processor): State: WAITING Blocked count: 1 Waited count: 350 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@42b32cdf Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.processQueue(BPServiceActor.java:1411) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor$CommandProcessingThread.run(BPServiceActor.java:1395) Thread 206 (BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527): State: TIMED_WAITING Blocked count: 1414 Waited count: 1615 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.IncrementalBlockReportManager.waitTillNextIBR(IncrementalBlockReportManager.java:158) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:771) app//org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:914) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 207 (pool-46-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 159 (org.apache.hadoop.hdfs.server.datanode.DataXceiverServer@459cf68): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.Net.accept(Native Method) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.implAccept(ServerSocketChannelImpl.java:425) java.base@17.0.11/sun.nio.ch.ServerSocketChannelImpl.accept(ServerSocketChannelImpl.java:391) java.base@17.0.11/sun.nio.ch.ServerSocketAdaptor.accept(ServerSocketAdaptor.java:126) app//org.apache.hadoop.hdfs.net.TcpPeerServer.accept(TcpPeerServer.java:85) app//org.apache.hadoop.hdfs.server.datanode.DataXceiverServer.run(DataXceiverServer.java:242) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 201 (IPC Server Responder): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.hadoop.ipc.Server$Responder.doRunLoop(Server.java:1733) app//org.apache.hadoop.ipc.Server$Responder.run(Server.java:1716) Thread 198 (IPC Server listener on 0): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hadoop.ipc.Server$Listener.run(Server.java:1559) Thread 208 (IPC Server handler 0 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 571 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 209 (IPC Server handler 1 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 571 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 210 (IPC Server handler 2 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 571 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 211 (IPC Server handler 3 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 574 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 212 (IPC Server handler 4 on default port 36113): State: TIMED_WAITING Blocked count: 0 Waited count: 571 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.poll(LinkedBlockingQueue.java:460) app//org.apache.hadoop.ipc.CallQueueManager.take(CallQueueManager.java:370) app//org.apache.hadoop.ipc.Server$Handler.run(Server.java:3165) Thread 215 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 216 (VolumeScannerThread(/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6)): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hdfs.server.datanode.VolumeScanner.run(VolumeScanner.java:656) Thread 222 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 221 (refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5/current/BP-1555733941-172.17.0.3-1736302801782): State: TIMED_WAITING Blocked count: 3 Waited count: 3 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.fs.CachingGetSpaceUsed$RefreshThread.run(CachingGetSpaceUsed.java:225) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 226 (pool-36-thread-1): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 229 (java.util.concurrent.ThreadPoolExecutor$Worker@3fdf04ce[State = -1, empty queue]): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 234 (FsDatasetAsyncDiskServiceFixer): State: TIMED_WAITING Blocked count: 0 Waited count: 20 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.HBaseTestingUtility$FsDatasetAsyncDiskServiceFixer.run(HBaseTestingUtility.java:599) Thread 237 (NIOServerCxnFactory.SelectorThread-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 236 (NIOServerCxnFactory.SelectorThread-0): State: RUNNABLE Blocked count: 14 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.select(NIOServerCnxnFactory.java:403) app//org.apache.zookeeper.server.NIOServerCnxnFactory$SelectorThread.run(NIOServerCnxnFactory.java:368) Thread 238 (NIOServerCxnFactory.AcceptThread:localhost/127.0.0.1:61593): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.select(NIOServerCnxnFactory.java:205) app//org.apache.zookeeper.server.NIOServerCnxnFactory$AcceptThread.run(NIOServerCnxnFactory.java:181) Thread 235 (ConnnectionExpirer): State: TIMED_WAITING Blocked count: 0 Waited count: 58 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.NIOServerCnxnFactory$ConnectionExpirerThread.run(NIOServerCnxnFactory.java:554) Thread 239 (SessionTracker): State: TIMED_WAITING Blocked count: 0 Waited count: 286 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.zookeeper.server.SessionTrackerImpl.run(SessionTrackerImpl.java:163) Thread 240 (SyncThread:0): State: WAITING Blocked count: 29 Waited count: 753 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@3adc2545 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.SyncRequestProcessor.run(SyncRequestProcessor.java:170) Thread 241 (ProcessThread(sid:0 cport:61593):): State: WAITING Blocked count: 1 Waited count: 862 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6eb1751a Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.PrepRequestProcessor.run(PrepRequestProcessor.java:142) Thread 242 (RequestThrottler): State: WAITING Blocked count: 0 Waited count: 895 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@11d419ff Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.server.RequestThrottler.run(RequestThrottler.java:147) Thread 243 (NIOWorkerThread-1): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 254 (Time-limited test.named-queue-events-pool-0): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34a48fd7 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) app//com.lmax.disruptor.BlockingWaitStrategy.waitFor(BlockingWaitStrategy.java:47) app//com.lmax.disruptor.ProcessingSequenceBarrier.waitFor(ProcessingSequenceBarrier.java:56) app//com.lmax.disruptor.BatchEventProcessor.processEvents(BatchEventProcessor.java:159) app//com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:125) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 255 (HBase-Metrics2-1): State: TIMED_WAITING Blocked count: 0 Waited count: 365 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 256 (RS-EventLoopGroup-1-1): State: RUNNABLE Blocked count: 41 Waited count: 2 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 257 (Time-limited test-SendThread(127.0.0.1:61593)): State: RUNNABLE Blocked count: 41 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:141) app//org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:332) app//org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1289) Thread 258 (Time-limited test-EventThread): State: WAITING Blocked count: 1 Waited count: 56 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@60f2d5c9 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:550) Thread 259 (NIOWorkerThread-2): State: WAITING Blocked count: 3 Waited count: 135 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 260 (NIOWorkerThread-3): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 261 (NIOWorkerThread-4): State: WAITING Blocked count: 3 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 262 (zk-event-processor-pool-0): State: WAITING Blocked count: 37 Waited count: 86 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@44149220 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 263 (NIOWorkerThread-5): State: WAITING Blocked count: 5 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 264 (NIOWorkerThread-6): State: WAITING Blocked count: 0 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 265 (NIOWorkerThread-7): State: WAITING Blocked count: 1 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 266 (NIOWorkerThread-8): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 267 (NIOWorkerThread-9): State: WAITING Blocked count: 1 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 268 (NIOWorkerThread-10): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 269 (NIOWorkerThread-11): State: WAITING Blocked count: 0 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 270 (NIOWorkerThread-12): State: WAITING Blocked count: 1 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 271 (NIOWorkerThread-13): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 272 (NIOWorkerThread-14): State: WAITING Blocked count: 3 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 273 (NIOWorkerThread-15): State: WAITING Blocked count: 2 Waited count: 133 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 274 (NIOWorkerThread-16): State: WAITING Blocked count: 2 Waited count: 134 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6f09797e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 275 (RpcServer.default.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@54de7742 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 276 (RpcServer.default.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 185 Waited count: 692 Waiting on java.util.concurrent.Semaphore$NonfairSync@44395ff0 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 277 (RpcServer.default.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 125 Waited count: 495 Waiting on java.util.concurrent.Semaphore$NonfairSync@32939623 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 278 (RpcServer.priority.RWQ.Fifo.write.handler=0,queue=0,port=41063): State: WAITING Blocked count: 90 Waited count: 4903 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@52f697c3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 279 (RpcServer.priority.RWQ.Fifo.read.handler=1,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 280 (RpcServer.priority.RWQ.Fifo.read.handler=2,queue=1,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@75359eb8 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.ipc.RpcHandler.getCallRunner(RpcHandler.java:71) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 281 (RpcServer.replication.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@7b7d9931 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 282 (RpcServer.replication.FPBQ.Fifo.handler=1,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@55713af3 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 283 (RpcServer.replication.FPBQ.Fifo.handler=2,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 1 Waiting on java.util.concurrent.Semaphore$NonfairSync@33dff711 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 284 (RpcServer.metaPriority.FPBQ.Fifo.handler=0,queue=0,port=41063): State: WAITING Blocked count: 0 Waited count: 3 Waiting on java.util.concurrent.Semaphore$NonfairSync@6a467f8c Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:211) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquire(AbstractQueuedSynchronizer.java:715) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1047) java.base@17.0.11/java.util.concurrent.Semaphore.acquire(Semaphore.java:318) app//org.apache.hadoop.hbase.ipc.FastPathRpcHandler.getCallRunner(FastPathRpcHandler.java:55) app//org.apache.hadoop.hbase.ipc.RpcHandler.run(RpcHandler.java:85) Thread 288 (RS-EventLoopGroup-3-1): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 308 (RS-EventLoopGroup-4-1): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 328 (RS-EventLoopGroup-5-1): State: RUNNABLE Blocked count: 90 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 286 (M:0;c4e5d52b9759:41063): State: TIMED_WAITING Blocked count: 6 Waited count: 2676 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.FutureTask.awaitDone(FutureTask.java:444) java.base@17.0.11/java.util.concurrent.FutureTask.get(FutureTask.java:203) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL.shutdown(AbstractFSWAL.java:1011) app//org.apache.hadoop.hbase.wal.AbstractFSWALProvider.shutdown(AbstractFSWALProvider.java:184) app//org.apache.hadoop.hbase.wal.WALFactory.shutdown(WALFactory.java:272) app//org.apache.hadoop.hbase.master.region.MasterRegion.shutdownWAL(MasterRegion.java:140) app//org.apache.hadoop.hbase.master.region.MasterRegion.close(MasterRegion.java:206) app//org.apache.hadoop.hbase.master.HMaster.stopServiceThreads(HMaster.java:1769) app//org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1285) app//org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:614) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 349 (Monitor thread for TaskMonitor): State: TIMED_WAITING Blocked count: 0 Waited count: 57 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.monitoring.TaskMonitor$MonitorRunnable.run(TaskMonitor.java:325) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 351 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 353 (master/c4e5d52b9759:0:becomeActiveMaster-MemStoreChunkPool Statistics): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 355 (org.apache.hadoop.hdfs.PeerCache@7a779c6a): State: TIMED_WAITING Blocked count: 0 Waited count: 190 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hdfs.PeerCache.run(PeerCache.java:253) app//org.apache.hadoop.hdfs.PeerCache.access$000(PeerCache.java:46) app//org.apache.hadoop.hdfs.PeerCache$1.run(PeerCache.java:124) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 374 (master:store-WAL-Roller): State: TIMED_WAITING Blocked count: 0 Waited count: 5666 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) app//org.apache.hadoop.hbase.wal.AbstractWALRoller.run(AbstractWALRoller.java:179) Thread 391 (RS-EventLoopGroup-5-2): State: RUNNABLE Blocked count: 78 Waited count: 3 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 392 (RS-EventLoopGroup-5-3): State: RUNNABLE Blocked count: 79 Waited count: 4 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 404 (Idle-Rpc-Conn-Sweeper-pool-0): State: WAITING Blocked count: 0 Waited count: 67 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@34577861 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 415 (SnapshotHandlerChoreCleaner): State: TIMED_WAITING Blocked count: 0 Waited count: 57 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 403 (RpcClient-timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 56594 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 424 (RS-EventLoopGroup-1-2): State: RUNNABLE Blocked count: 42 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 425 (RS-EventLoopGroup-1-3): State: RUNNABLE Blocked count: 34 Waited count: 1 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 450 (RegionServerTracker-0): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@387eb41e Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 472 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 13 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7303e525 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 473 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 17 Waited count: 27 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@7c180f54 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 474 (regionserver/c4e5d52b9759:0.procedureResultReporter): State: WAITING Blocked count: 18 Waited count: 31 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6dc8d2bc Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) app//org.apache.hadoop.hbase.regionserver.RemoteProcedureResultReporter.run(RemoteProcedureResultReporter.java:75) Thread 512 (RS-EventLoopGroup-4-2): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 523 (RS-EventLoopGroup-4-3): State: RUNNABLE Blocked count: 2 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 525 (region-location-0): State: WAITING Blocked count: 9 Waited count: 15 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 554 (Async-Client-Retry-Timer-pool-0): State: TIMED_WAITING Blocked count: 0 Waited count: 56374 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.waitForNextTick(HashedWheelTimer.java:598) app//org.apache.hbase.thirdparty.io.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:494) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 565 (RPCClient-NioEventLoopGroup-6-1): State: RUNNABLE Blocked count: 5 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 594 (RS-EventLoopGroup-3-2): State: RUNNABLE Blocked count: 1 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 606 (region-location-1): State: WAITING Blocked count: 1 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 607 (region-location-2): State: WAITING Blocked count: 2 Waited count: 6 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 608 (region-location-3): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1007 (MutableQuantiles-0): State: TIMED_WAITING Blocked count: 0 Waited count: 422 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1182) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1073 (RPCClient-NioEventLoopGroup-6-2): State: RUNNABLE Blocked count: 6 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1109 (RS-EventLoopGroup-3-3): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native Method) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:220) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.Native.epollWait(Native.java:213) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.epollWaitNoTimerChange(EpollEventLoop.java:308) app//org.apache.hbase.thirdparty.io.netty.channel.epoll.EpollEventLoop.run(EpollEventLoop.java:365) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1111 (zk-permission-watcher-pool-0): State: WAITING Blocked count: 60 Waited count: 96 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5bb51352 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1119 (RPCClient-NioEventLoopGroup-6-3): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 1566 (Container metrics unregistration): State: WAITING Blocked count: 10 Waited count: 33 Waiting on java.util.TaskQueue@213e44c6 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.lang.Object.wait(Object.java:338) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:537) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 2096 (RPCClient-NioEventLoopGroup-6-4): State: RUNNABLE Blocked count: 3 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 2998 (region-location-4): State: WAITING Blocked count: 2 Waited count: 5 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@20db455f Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1170) java.base@17.0.11/java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:899) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5090 (RPCClient-NioEventLoopGroup-6-5): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5091 (RPCClient-NioEventLoopGroup-6-6): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 5092 (RPCClient-NioEventLoopGroup-6-7): State: RUNNABLE Blocked count: 0 Waited count: 0 Stack: java.base@17.0.11/sun.nio.ch.EPoll.wait(Native Method) java.base@17.0.11/sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:118) java.base@17.0.11/sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:129) java.base@17.0.11/sun.nio.ch.SelectorImpl.select(SelectorImpl.java:146) app//org.apache.hbase.thirdparty.io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:68) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:879) app//org.apache.hbase.thirdparty.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:526) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) app//org.apache.hbase.thirdparty.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) app//org.apache.hbase.thirdparty.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9177 (AsyncFSWAL-1-hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData-prefix:c4e5d52b9759,41063,1736302807767): State: WAITING Blocked count: 0 Waited count: 2 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@37ae9b Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.park(LockSupport.java:341) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionNode.block(AbstractQueuedSynchronizer.java:506) java.base@17.0.11/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3465) java.base@17.0.11/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3436) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1625) java.base@17.0.11/java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:435) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1062) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9181 (Timer for 'JobHistoryServer' metrics system): State: TIMED_WAITING Blocked count: 0 Waited count: 26 Stack: java.base@17.0.11/java.lang.Object.wait(Native Method) java.base@17.0.11/java.util.TimerThread.mainLoop(Timer.java:563) java.base@17.0.11/java.util.TimerThread.run(Timer.java:516) Thread 9182 (process reaper): State: TIMED_WAITING Blocked count: 0 Waited count: 1 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.SynchronousQueue$TransferStack.transfer(SynchronousQueue.java:401) java.base@17.0.11/java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:903) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1061) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1122) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9184 (WAL-Shutdown-0): State: TIMED_WAITING Blocked count: 0 Waited count: 2 Stack: java.base@17.0.11/jdk.internal.misc.Unsafe.park(Native Method) java.base@17.0.11/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) java.base@17.0.11/java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.awaitNanos(AbstractQueuedSynchronizer.java:1674) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.awaitTermination(ThreadPoolExecutor.java:1464) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.doShutdown(AsyncFSWAL.java:793) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$2.call(AbstractFSWAL.java:995) app//org.apache.hadoop.hbase.regionserver.wal.AbstractFSWAL$2.call(AbstractFSWAL.java:990) java.base@17.0.11/java.util.concurrent.FutureTask.run(FutureTask.java:264) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) Thread 9185 (Close-WAL-Writer-0): State: TIMED_WAITING Blocked count: 1 Waited count: 2 Stack: java.base@17.0.11/java.lang.Thread.sleep(Native Method) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:166) app//org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) app//org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL$$Lambda$1123/0x00007ff381163030.run(Unknown Source) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) java.base@17.0.11/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) java.base@17.0.11/java.lang.Thread.run(Thread.java:840) 2025-01-08T02:29:40,400 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(233): attempt=1 on file=hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644 after 4001ms java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.recoverLease(DFSClient.java:946) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:317) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$3.doCall(DistributedFileSystem.java:314) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.recoverLease(DistributedFileSystem.java:329) ~[hadoop-hdfs-client-3.4.1.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverLease(RecoverLeaseFSUtils.java:222) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:155) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] 2025-01-08T02:29:41,396 ERROR [WAL-Shutdown-0 {}] wal.AsyncFSWAL(794): We have waited 5 seconds but the close of async writer doesn't complete.Please check the status of underlying filesystem or increase the wait time by the config "hbase.wal.async.wait.on.shutdown.seconds" 2025-01-08T02:29:41,396 INFO [master:store-WAL-Roller {}] wal.AbstractWALRoller(243): LogRoller exiting. 2025-01-08T02:29:41,396 INFO [M:0;c4e5d52b9759:41063 {}] flush.MasterFlushTableProcedureManager(91): stop: server shutting down. 2025-01-08T02:29:41,397 INFO [M:0;c4e5d52b9759:41063 {}] ipc.NettyRpcServer(351): Stopping server on /172.17.0.3:41063 2025-01-08T02:29:41,398 DEBUG [M:0;c4e5d52b9759:41063 {}] zookeeper.RecoverableZooKeeper(215): Node /hbase/rs/c4e5d52b9759,41063,1736302807767 already deleted, retry=false 2025-01-08T02:29:41,402 WARN [Close-WAL-Writer-0 {}] util.RecoverLeaseFSUtils(258): Failed invocation for hdfs://localhost:43527/user/jenkins/test-data/14937bd5-485f-14cb-1976-edf92e0ce572/MasterData/WALs/c4e5d52b9759,41063,1736302807767/c4e5d52b9759%2C41063%2C1736302807767.1736302809644 java.lang.reflect.InvocationTargetException: null at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.isFileClosed(RecoverLeaseFSUtils.java:254) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverDFSFileLease(RecoverLeaseFSUtils.java:186) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.util.RecoverLeaseFSUtils.recoverFileLease(RecoverLeaseFSUtils.java:96) ~[hbase-asyncfs-2.7.0-SNAPSHOT.jar:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.recoverLease(AsyncFSWAL.java:722) ~[classes/:?] at org.apache.hadoop.hbase.regionserver.wal.AsyncFSWAL.lambda$closeWriter$5(AsyncFSWAL.java:735) ~[classes/:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?] at java.lang.Thread.run(Thread.java:840) ~[?:?] Caused by: java.io.IOException: Filesystem closed at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:490) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DFSClient.isFileClosed(DFSClient.java:1808) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2539) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem$46.doCall(DistributedFileSystem.java:2536) ~[hadoop-hdfs-client-3.4.1.jar:?] at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) ~[hadoop-common-3.4.1.jar:?] at org.apache.hadoop.hdfs.DistributedFileSystem.isFileClosed(DistributedFileSystem.java:2554) ~[hadoop-hdfs-client-3.4.1.jar:?] ... 12 more 2025-01-08T02:29:41,500 INFO [M:0;c4e5d52b9759:41063 {}] regionserver.HRegionServer(1307): Exiting; stopping=c4e5d52b9759,41063,1736302807767; zookeeper connection closed. 2025-01-08T02:29:41,500 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:29:41,500 DEBUG [Time-limited test-EventThread {}] zookeeper.ZKWatcher(609): master:41063-0x1015754588a0000, quorum=127.0.0.1:61593, baseZNode=/hbase Received ZooKeeper Event, type=None, state=Closed, path=null 2025-01-08T02:29:41,503 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@12bc8767{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T02:29:41,504 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@e141426{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:29:41,504 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:29:41,504 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@178a4d2d{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T02:29:41,504 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@1887d910{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:29:41,505 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T02:29:41,505 WARN [BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T02:29:41,505 WARN [BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1555733941-172.17.0.3-1736302801782 (Datanode Uuid 5f73ff73-0f7b-4bdf-9032-a33a82e289db) service to localhost/127.0.0.1:43527 2025-01-08T02:29:41,505 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T02:29:41,507 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data5/current/BP-1555733941-172.17.0.3-1736302801782 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T02:29:41,507 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data6/current/BP-1555733941-172.17.0.3-1736302801782 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T02:29:41,507 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T02:29:41,509 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@6e8b8a6d{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T02:29:41,510 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@62d7732{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:29:41,510 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:29:41,510 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@24e59c37{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T02:29:41,510 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@64a3ece4{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:29:41,512 WARN [BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T02:29:41,512 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T02:29:41,512 WARN [BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1555733941-172.17.0.3-1736302801782 (Datanode Uuid 66baa9c9-8a2d-42eb-9f59-398f6cb5787c) service to localhost/127.0.0.1:43527 2025-01-08T02:29:41,512 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T02:29:41,512 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data3/current/BP-1555733941-172.17.0.3-1736302801782 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T02:29:41,512 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data4/current/BP-1555733941-172.17.0.3-1736302801782 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T02:29:41,513 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T02:29:41,517 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@5a7079af{datanode,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/datanode} 2025-01-08T02:29:41,517 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@76ce3a96{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:29:41,517 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:29:41,517 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@3299e9d3{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T02:29:41,518 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@74ef64a6{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:29:41,519 WARN [BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527 {}] datanode.IncrementalBlockReportManager(160): IncrementalBlockReportManager interrupted 2025-01-08T02:29:41,519 WARN [BP-1555733941-172.17.0.3-1736302801782 heartbeating to localhost/127.0.0.1:43527 {}] datanode.BPServiceActor(925): Ending block pool service for: Block pool BP-1555733941-172.17.0.3-1736302801782 (Datanode Uuid d1c88fec-0d29-4195-9b79-b97d3d006240) service to localhost/127.0.0.1:43527 2025-01-08T02:29:41,519 ERROR [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1416): Command processor encountered interrupt and exit. 2025-01-08T02:29:41,519 WARN [Command processor {}] datanode.BPServiceActor$CommandProcessingThread(1400): Ending command processor service for: Thread[Command processor,5,FailOnTimeoutGroup] 2025-01-08T02:29:41,519 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data1/current/BP-1555733941-172.17.0.3-1736302801782 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T02:29:41,519 WARN [refreshUsed-/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/cluster_00f5710b-3e42-f65d-b988-500e94518f56/dfs/data/data2/current/BP-1555733941-172.17.0.3-1736302801782 {}] fs.CachingGetSpaceUsed$RefreshThread(231): Thread Interrupted waiting to refresh disk information: sleep interrupted 2025-01-08T02:29:41,520 WARN [Time-limited test {}] datanode.DataSetLockManager(261): not open lock leak check func 2025-01-08T02:29:41,526 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.w.WebAppContext@603d2b6a{hdfs,/,null,STOPPED}{jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/hdfs} 2025-01-08T02:29:41,527 INFO [Time-limited test {}] server.AbstractConnector(383): Stopped ServerConnector@3797254c{HTTP/1.1, (http/1.1)}{localhost:0} 2025-01-08T02:29:41,527 INFO [Time-limited test {}] session.HouseKeeper(149): node0 Stopped scavenging 2025-01-08T02:29:41,527 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@6f0daeb1{static,/static,jar:file:/home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/local-repository/org/apache/hadoop/hadoop-hdfs/3.4.1/hadoop-hdfs-3.4.1-tests.jar!/webapps/static,STOPPED} 2025-01-08T02:29:41,527 INFO [Time-limited test {}] handler.ContextHandler(1159): Stopped o.e.j.s.ServletContextHandler@690f3f23{logs,/logs,file:///home/jenkins/jenkins-home/workspace/HBase-Flaky-Tests_branch-2/hbase-mapreduce/target/test-data/27ea3367-5c8d-7204-cd4a-9676169d75bf/hadoop.log.dir/,STOPPED} 2025-01-08T02:29:41,541 INFO [Time-limited test {}] zookeeper.MiniZooKeeperCluster(345): Shutdown MiniZK cluster with all ZK servers 2025-01-08T02:29:41,788 INFO [Time-limited test {}] hbase.HBaseTestingUtility(1347): Minicluster is down